Multi-Modal Language Tutor Using GPT-4 Turbo, Whisper, and OpenAI Text-to-Speech
Offered By: echohive via YouTube
Course Description
Overview
Explore a multi-modal language tutor that leverages GPT-4 Turbo, Whisper, and OpenAI Text to Speech to translate requests between multiple languages automatically from voice or text input. Learn how responses are saved to structured markdown files and audio responses are stored as MP3 files for future review. Dive into the main loop code, understand the importance of threading, and see demonstrations of the tutor in action. Discover how to switch between voice and text modes, get situational language help, and utilize the Markdown preview extension in VS Code. Gain insights into the CodeHive collection of 900+ free GPT Python apps and explore additional resources for mastering GPT API and voice-controlled Auto AGI capabilities.
Syllabus
INTRO AND DEMO
Conversation saved to markdown file
Mode switchingbetween Voice and Text
Situational language help
Main loop code review
Markdown preview extension in VS Code
CodeHive 900+ Free GPT python apps
Necessity of using Threading
Full Code review
More Demos
Taught by
echohive
Related Courses
Building AI Applications with Watson APIsIBM via Coursera Microsoft Cognitive Services: Azure Custom Text to Speech
Pluralsight Getting Started with Xamarin.Essentials in Xamarin.Forms
Pluralsight Learning Microsoft Cognitive Services for Developers
LinkedIn Learning Microsoft Cognitive Services for Developers: 2 Speech
LinkedIn Learning