This study explores the use of machine learning models to predict depression through the analysis of verbal and non-verbal cues, highlighting the integration of multimodal data including voice and facial expressions. Findings indicate that the proposed model achieved an accuracy of 83%, suggesting its potential as a reliable diagnostic tool compared to traditional assessments. The research underscores the importance of comprehensive data collection and a multimodal approach for improving depression prediction in mental health care.