Skip to content

Madhudeepika/DataScience_Projects

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

34 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Data Science Projects

Project 01 : Medical Inventory Optimization & Forecasting

Project Description: This project focuses on addressing the notable surge in bounce rates within the healthcare system, adversely affecting patient satisfaction. The aim is to optimize medical inventory levels to minimize bounce rates while keeping inventory costs at a minimum.

Key Features:

  • Data Preprocessing: Explore and preprocess historical data to ensure accuracy in forecasting.

  • SARIMAX Model Implementation: Implement SARIMAX models for future bounce rate predictions.

  • Streamlit Model Deployment: Utilize Streamlit for real-time processing and transparent display of forecasted values and dates.

  • Data Visualization: Enhance comprehension with line charts and histograms to facilitate effective decision-making.

  • Continuous Monitoring: Ensure continuous monitoring and feedback for iterative improvements and validation of strategies.

Objective: The primary objective of this project is to achieve a minimum 30% reduction in bounce rates and drive a revenue increase of at least 20 lacs INR through effective inventory optimization.

Project Scope:

  • Analyze historical bounce rates and key contributing factors.

  • Explore and preprocess relevant data for accurate forecasting.

  • Implement SARIMAX models for future bounce rate predictions.

  • Optimize inventory levels to align with forecasted demand.

  • Enhance patient experience through streamlined operations.

Conclusion: The forecasted values provide promising insights into the effectiveness of the implemented strategies in minimizing bounce rates and optimizing inventory costs. Continuous monitoring and assessment will be crucial to validate the actual impact on bounce rates and economic success. Further data analysis and stakeholder feedback will enhance the depth of our understanding and validate the success of the implemented strategies.

Project 02 : Liver Tumor classification using Deep Learning

Contents:

  • Business Problem
  • Project Overview and Scope
  • Data Collection & Data Preprocessing
  • Data Spliting
  • Project Architecture
  • Model Building
  • Model Deployment strategy
  • Final Output
  • Conclusion
  • Challenges
  • Future Work

Business Problem:

  • Business Problem:

    Bounce rate is increasing significantly leading to patient dissatisfaction.
  • Business Objective:

    Minimize bounce rate
  • Business Constraint:

    Minimize inventory cost
  • Success Criteria:

  • Business Success Criteria:
    Reduce bounce rate by at least 30%
  • Economic Success Criteria
    Increase revenue by at least 20 lacs INR by reducing bounce rate

Project Overview and Scope

Project Goals and Objectives:

  • Develop a deep Learning model for liver tumor classification.
  • Minimize Bounce rate in medical diagnosis by improving accuracy and efficiency.
  • Enhance Patient Satisfaction through accurate and timely diagnosis.

Scope of the Project:

  • Utilize deep learning techniques to analyze medical images for liver tumor detection.
  • Implement data preprocessing,model building and deployment phases.
  • Evaluate model performance on reducing bounce rate and improving patient outcomes.
  • Address the business problem by providing a scalable solution for accurate tumor classification.
  • Collaborate with medical professionals to ensure clinical relevance and effectiveness of the model.

Data Collection & Data Preprocessing:

  • The Data was collected from the primary source: Directly Received from the Client. DataSet was Histopathological images of Liver tumors and Normal liver.

Data Preprocessing:

Techniques Used for Data Cleaning and Augmentation:
  • Removal of irrelevant or noisy data points.
  • Standardization of image sizes and resolutions.
  • Augmentation techniques such as rotation, flipping and zooming to increase dataset diversity.
Importance of Data Preprocessing for Model Training:
  • Enhances model generalization by reducing overfitting.
  • Improves Model robustness to variations in input data.
  • Ensures quality and consistency of input data for reliable model predictions.

Data Splitting

Explanation of Data split into Training and Testing Sets:
  • Divide the dataset into two subsets: Training and Testing
  • Training set: Used to train the deep learning model.
  • Testing Set: Used to Evaluate the Model's Performance on unseen data.
Ensuring Model Generalization with Proper Data Splitting:
  • Random selection ensures representative samples in both sets.
  • Typically a split of 70-80% for training and 20-30% for testing is used.
  • Helps assess how well the model will perform on new, unseen data.
  • Prevents over fitting by validating model performance on independent data.

Project Architecture

Project Architecture

Model Building:

Description of Attempted Models:
  • DenseNet121 , ResNet50 , ResNet152, CNN , Vgg16, Yolov8
  • Each model has its unique architecture and capabilities.
  • Tried various models to determine the most suitable for live tumor classification.
Selection Criteria and Rationale for Choosing ResNet50:
  • ResNet50 demonstrated a balance between accuracy and computational efficiency.
  • Known for its deep architecture and skip connections, allowing for effective feature extraction.
  • Widely used and well documented, making it easier for implementation and troubleshooting.
  • Achieved satisfactory performance during preliminary testing.

ResNet50 Model:

Detailed Architecture of ResNet50:
  • Deep Neural Network: ResNet50 Comprises 50 layers, allowing for deep learning capabilities
  • Skip Connections: If employs skip connections to alleviate the vanishing gradient problem, facilitating the training of deeper networks.
  • Residual Blocks: Built using residual blocks that add the input to the output, enabling the learning of residual functions.
  • Bottleneck Layers: Utilizes bottleneck layers to reduce computational complexity while maintaining model depth, enhancing efficiency.

Training Process and Achieved Accuracy (96.5%)

Training Process:
  • Data Preprocessing:The dataset underwent preprocessing steps such as resizing , normalization and augmentation to enhance model performance and generalization.
  • Hyperparameter Tuning: Hyperparameters like learning rate, batch size and optimizer choice were tuned to optimize model convergence and accuracy.
  • Optimization techniques: Techniques like Stochastic gradient Descent (SGD), Adam , pr RMSprop were utilized to minimize the loss function during training
Achieved Accuracy:
  • After training on the preprocessed dataset, the ResNet50 model achieved an accuracy of 96.5% on the testing dataset.
  • This high accuracy demonstrates the effectiveness of the model in accurately classifying liver tumors, thereby contributing to the projects objectives of reducing bounce rates and improving patient satisfaction in medical diagnosis

Deployment

Deployment of ResNet50 using Flask:
  • Flask Framework: Utilized Flask, A lightweight web framework for deploying the ResNet50 model.
Steps Involved:
  • Integration: Incorporate the trained ResNet50 model into a Flask web application
  • Endpoint Creation: define endpoints to receive image data for classification
  • Prediction: Implement logic to process incoming data and generate predictions using the ResNet50 Model.
  • Response Handling: Send classification results back to the user interface for display.

Final Output

Finaloutput

Conclusion

Summary of Project Outcomes and impact on Business objectives
  • The Project Successfully deployed the ResNet50 model for liver tumor classification , aligning with the objective of minimizing bounce rates in medical diagnosis. By accurately identifying liver tumors the model contributes to improved patient satisfaction and healthcare outcomes.
Importance of Deep Learning in medical Diagnostics:
  • Deep Learning techniques, exemplified by ResNet50 ,play a pivotal role in medical diagnostics.
  • They enable precise analysis of medical images, enhancing diagnostic accuracy and facilitating timely treatment decisions.

Challenges

Challenges Encountered During Project Implementation:
  • Data preprocessing complexities, including normalization and augmentation posed initial challenges.
  • Optimizing hyper parameters and managing computational resources were key hurdles during model training.
Strategies Employed to Address challenges and Ensure Success:
  • Collaboration with domain experts helped refine preprocessing techniques and ensure clinical relevance.
  • Fine-tuning hyper parameters and utilizing cloud computing resources addressed computational constraints leading to successful model training and deployment

Future Work

Potential Advancements in Deep Learning for medical image Analysis:
  • Emerging trends like multi-modal fusion and transfer learning hold promise for advancing medical image analysis.
  • Continued research in areas like explainable AI and uncertainity estimation can further enhance the reliability and interpretability of deep learning models in medical diagnostics.

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors