Multi-Modal Language Tutor Using GPT-4 Turbo, Whisper, and OpenAI Text-to-Speech
Offered By: echohive via YouTube
Course Description
Overview
Explore a multi-modal language tutor that leverages GPT-4 Turbo, Whisper, and OpenAI Text to Speech to translate requests between multiple languages automatically from voice or text input. Learn how responses are saved to structured markdown files and audio responses are stored as MP3 files for future review. Dive into the main loop code, understand the importance of threading, and see demonstrations of the tutor in action. Discover how to switch between voice and text modes, get situational language help, and utilize the Markdown preview extension in VS Code. Gain insights into the CodeHive collection of 900+ free GPT Python apps and explore additional resources for mastering GPT API and voice-controlled Auto AGI capabilities.
Syllabus
INTRO AND DEMO
Conversation saved to markdown file
Mode switchingbetween Voice and Text
Situational language help
Main loop code review
Markdown preview extension in VS Code
CodeHive 900+ Free GPT python apps
Necessity of using Threading
Full Code review
More Demos
Taught by
echohive
Related Courses
Machine Learning Capstone: An Intelligent Application with Deep LearningUniversity of Washington via Coursera Elaborazione del linguaggio naturale
University of Naples Federico II via Federica Deep Learning for Natural Language Processing
University of Oxford via Independent Deep Learning Summer School
Independent Sequence Models
DeepLearning.AI via Coursera