Date: February 21, 2025
The Voice-to-Action App is a React Native application that allows users to record audio, transcribe it into text, and extract actionable insights such as meeting details, key points, decisions, and follow-up tasks. The app integrates with Google Speech-to-Text API for transcription and Perplexity AI for advanced analysis of the transcript.
This project is designed to streamline meeting management by providing users with:
- A transcript of the recording.
- Structured meeting summaries.
- Actionable to-do lists.
- Email integration for sharing meeting details.
- Record audio directly within the app using the device's microphone.
- Supports both iOS and Android platforms.
- Converts recorded audio into text using the Google Speech-to-Text API.
- Extracts structured data from transcripts using Perplexity AI, including:
- Meeting details (date, time, participants).
- Key discussion points.
- Decisions made.
- Action items with assignees and deadlines.
- Compose and send meeting summaries via email directly from the app.
- Modify meeting details and tasks through an intuitive edit modal.
- A visually appealing dark mode design for better user experience.
Here is my main link where I have helped people understand how my project works by giving a hands-on demo:
Watch Demo
Here is my second link where I have explained how the code works in the backend and frontend, line by line:
Watch Code Walkthrough
- React Native
- Expo AV (for audio recording)
- Axios (for backend communication)
- React Native Modal (for modals)
- React Native Linking (for email integration)
- Node.js with Express.js
- Google Speech-to-Text API
- Perplexity AI API
- Install Node.js and npm.
- Install Expo CLI for running the React Native app.
- Set up a Google Cloud account and enable the Speech-to-Text API.
- Obtain an API key from Perplexity AI.
- Clone the repository and navigate to the backend folder:
git clone cd backend
- Install dependencies:
npm install
- Add your Google Cloud credentials JSON file (
myjson.json
) to the backend folder. - Replace
perplexityApikey
intranscriptionService.js
with your Perplexity AI API key. - Start the server:
node index.js
- Navigate to the frontend folder:
cd frontend
- Install dependencies:
npm install
- Update the
myip
variable inApp.js
with your local machine's IP address (e.g.,192.168.x.x
). - Start the Expo development server:
expo start
- Scan the QR code from Expo on your mobile device or run it on an emulator.
- Launch the app on your mobile device or emulator.
- Tap "🎤 Start Recording" to begin recording audio.
- Tap "⏹️ Stop Recording" to stop recording and process the audio.
- View the generated transcript and structured insights (meeting details, tasks, etc.).
- Edit any details by tapping the edit icon ✏️.
- Share meeting summaries via email by tapping the email icon 📧.
project/
├── backend/
│ ├── index.js # Main server file
│ ├── transcriptionService.js # Handles transcription and AI analysis
│ └── myjson.json # Google Cloud credentials file (not included in repo)
├── frontend/
│ ├── App.js # Main React Native app file
│ ├── styles.js # Styling for UI components
│ └── assets/ # App assets (if any)
└── README.md # Documentation file
Home Screen | Transcript & Meeting Details |
---|---|
![]() |
![]() |
- Add support for multiple languages in transcription.
- Integrate calendar APIs (e.g., Google Calendar) for automatic event creation.
- Enable offline transcription using local models.
For any questions or feedback, feel free to reach out:
Yasharth Bajpai
Email: yasharthbajpai0103@gmail.com