Giter Club home page Giter Club logo

barkamol_portfolio's Introduction

Barkamol_Portfolio

Barkamol Urinboev

Overview

This project aims to measure the credit risk of LendingClub, (an American peer-to-peer lending company), by calculating the expected loss of their outstanding loans. Credit risk is the likelihood that a borrower would not repay their loan to the lender. By continually evaluating the risk and adjusting their credit policies, the lender could minimize its credit losses while it reaches the fullest potential to maximize revenues on loan borrowing. It is also crucial for the lender to abide by regulations that require them to conduct their business with sufficient capital adequacy, which, if in low, will risk the stability of the economic system.

The key metric of credit risk is Expected Loss (EL), calculated by multiplying the results across three models: PD (Probability of Default), LGD (Loss Given Default), and EAD (Exposure at Default). The project includes all three models to help reach the final goal of credit risk measurement.

Requirements

Overview

This project aims to support a retail or FMCG (fast-moving consumer goods) company to formulate marketing and pricing strategies that could maximize revenues on each brand of candy bars. To reach the fullest potential of bringing up revenues, a company should find the 'sweet spot' for price to maximize three customer behaviours: purchase probability, brand choice probability, and purchase quantity.

Data from customer purchase history were used for training the regression models to predict those three customer behaviours in a preconceived price range. The results were then converted into price elasticities so that we can examine the effects of changing price on each of the behaviours. Hence, we will be able to find the suitable marketing and pricing strategies.

To better position our products, we will firstly perform segmentation on our customers to support our analysis on customer behaviours, allowing us to customize marketing strategies for customers with different backgrounds.

Requirements

Overview

This portfolio project is an analysis of a dataset on absenteeism in a company. The dataset includes various attributes, such as the reason for absence, the date of absence, the transportation expense, and others, for each employee. The project aims to identify the most common reasons for employee absenteeism, and to build a predictive model that can accurately predict whether an employee is likely to be absent or not.

The analysis of the data revealed that the most common reasons for absenteeism were related to diseases and medical appointments, followed by various other personal reasons, and then work-related reasons. Further analysis showed that certain factors, such as age, distance from work, and work load, were significant predictors of absenteeism.

To build a predictive model, the dataset was split into a training set and a test set. Different machine learning algorithms, including logistic regression, decision tree, and random forest, were applied to the training set, and their performance was evaluated using various metrics, such as accuracy, precision, recall, and F1-score.

The best-performing model was the random forest algorithm, which achieved an accuracy of 80% on the test set. The model was also able to identify the most important predictors of absenteeism, such as the reason for absence, the day of the week, and the age of the employee.

In conclusion, this project demonstrates the usefulness of data analysis and machine learning in identifying patterns and predicting outcomes in employee absenteeism. The insights and models developed in this project can be used by managers and HR professionals to better understand and manage employee absenteeism in their companies.

Requirements

Overview

The Tableau Portfolio Project is a collection of SQL queries and Tableau dashboards that explore a dataset of customer orders for a fictitious company. The dataset includes information on customer orders, products, and sales, and the goal of this project is to provide insights into the company's sales and customer behavior.

The SQL queries are used to prepare the data for analysis, and the results are then visualized in Tableau dashboards. The dashboards provide an interactive way to explore the data and to discover insights into the company's sales trends, customer behavior, and product performance.

The project demonstrates skills and knowledge in SQL, data cleaning and preparation, data analysis, and data visualization using Tableau. By exploring the data and creating interactive dashboards, this project provides valuable insights into the company's sales and customer behavior, which could be used to inform business decisions and strategies.

Requirements

Overview

The Tableau Public profile of "barkamol" showcases a collection of data visualizations and dashboards created by the author. The visualizations cover a variety of topics, including finance, marketing, and sports. The author uses Tableau's powerful visualization tools to create interactive dashboards that allow users to explore and analyze data in an intuitive and engaging way. The visualizations incorporate various chart types such as bar charts, scatterplots, and maps to effectively communicate insights and trends in the data. The profile provides valuable insights into the author's data visualization skills and demonstrates their ability to create compelling and informative data stories using Tableau.

Requirement

  • Packages: tableau

Overview

Data cleaning is a crucial step in the data analysis process, and it involves identifying and correcting errors, inconsistencies, and inaccuracies in data. The SQL script in this project provides a set of queries that can be used to perform various data cleaning tasks.

The script is well-organized and divided into sections that correspond to specific data cleaning tasks. For example, the section on handling missing data provides queries for identifying and handling missing values in a dataset. Similarly, the section on data type conversions provides queries for converting data from one data type to another.

One of the strengths of this project is that it covers a wide range of data cleaning tasks. The queries provided can be used to remove duplicates, standardize data, and address inconsistencies in data. The queries are also customizable, allowing users to modify them to suit their specific needs.

Overall, the "Data Cleaning SQL Queries" project is a valuable resource for anyone working with data in SQL. The queries provided can save time and effort by automating many data cleaning tasks and ensuring that data is accurate and ready for analysis. The project is a testament to the importance of data cleaning and the power of SQL as a tool for managing and manipulating data.

Requirements

Overview

How do data workers make money, which continent earns more money, are they satisfied with the money they are being paid, what programming language do they use, what is their level of data complexity, what is their life-work balance, and similar questions. If you want to find an answer, I recommend you to refer to this dashboard.

Requirement

  • Packages: power bi

Overview

The project is titled "COVID Portfolio Project - Data Exploration" and the code is written in SQL language. The main objective of the project is to explore and analyze COVID-19 data using various SQL queries and visualizations.

The project begins with importing COVID-19 data into a SQL database, followed by data cleaning and pre-processing. Then, the author uses various SQL queries to answer questions related to COVID-19, such as:

  • What is the total number of confirmed cases and deaths in each country?
  • How has the number of cases and deaths changed over time?
  • What are the top 10 countries with the highest number of cases and deaths?
  • How has the spread of COVID-19 varied by continent?

The author also uses SQL to create visualizations such as bar charts, line graphs, and heatmaps to help better understand the data.

Overall, the project provides a good example of how SQL can be used to analyze and visualize COVID-19 data, and may be useful for anyone interested in data exploration or epidemiology.

Requirements

Overview

The Resume Parser with NLP is a program that uses natural language processing (NLP) to extract essential information from resumes. The program receives resumes in PDF format, and it processes them using several NLP techniques to extract personal and professional details such as name, email, phone number, work experience, education, skills, and other relevant information.

The program uses several Python libraries such as spaCy, PyPDF2, and Regex to perform NLP tasks such as tokenization, parsing, named-entity recognition, and information extraction. It also uses machine learning models to classify the information and extract specific details.

The program is designed to automate the resume screening process, making it more efficient and accurate. It can be used by HR departments or recruitment agencies to quickly identify the most promising candidates and filter out irrelevant applications.

Overall, the Resume Parser with NLP is a powerful tool that streamlines the recruitment process and helps organizations find and hire the best candidates for their open positions.

Requirements

Overview

Which gender rides a bicycle more often, which continent, and whether they are married or not, how much it depends on the level of education (bachelor etc). We can find answers to these and similar questions through this dashboard.

Requirement

  • Packages: excel

Overview

The "Developer ChatBot with ChatGPT API" is a project that aims to develop a chatbot using the ChatGPT API to assist developers with their programming-related queries. The project is developed in a Jupyter Notebook, which allows for easy implementation and execution of the code.

The chatbot is implemented using a while loop, which keeps running until the user types "bye" to end the conversation. The ChatGPT API is used to generate the responses of the chatbot. To use the API, you need to send a POST request to the API endpoint with the user's input as the query parameter. The API returns a JSON object containing the response of the chatbot, which is then displayed to the user.

The project is a valuable tool for any programming community as it provides instant solutions to users' queries. It can be customized to suit different requirements, making it an excellent tool for developers who want to create a chatbot to help other developers. Overall, the "Developer ChatBot with ChatGPT API" project is an excellent example of how machine learning can be used to develop intelligent chatbots to assist users with their queries.

Requirements

  • Python Version: 3.9.6
  • Packages: gradio, and openai
  • Algorithms: OpenAI's GPT-3

Overview

Overview of Time Series Forecasting using FBProphet.ipynb:

The goal of this project is to demonstrate time series forecasting using the FBProphet library in Python. The dataset used in the project is the daily number of passengers for an airline company, spanning from 2018 to 2020. The project involves data preprocessing, visualization, modeling, forecasting, and evaluation of the model.

In data preprocessing, the "Month" column is parsed, the columns are renamed, and the "Passengers" column is converted to a numeric type. In data visualization, the time series data is plotted to visualize the trend and seasonality.

In modeling, a Prophet model is created and fit to the data. The model is trained on the first 80% of the data and tested on the remaining 20%. In forecasting, the model is used to forecast the number of passengers for the next few years.

In evaluation, the performance of the model is evaluated using various metrics such as mean absolute error, mean squared error, and root mean squared error. In visualization, the forecasted data is plotted along with the historical data to visualize the accuracy of the model.

The project demonstrates that FBProphet can be used to effectively forecast time series data. The model accurately captures the trend and seasonality of the data and provides valuable insights for decision-making. The notebook can be used as a guide for those interested in learning how to use FBProphet for time series forecasting.

Requirements

Overview

The project titled "Data Cleaning Portfolio Project Queries" is a SQL-based project that focuses on data cleaning and data manipulation. The project begins by importing data into a SQL database, followed by data cleaning and pre-processing using various SQL queries.

The author employs a wide range of SQL queries to clean and manipulate the data, such as:

  • Removing duplicate rows from the data.
  • Handling missing values using various techniques such as imputation and deletion.
  • Renaming columns and changing data types to improve data quality.
  • Normalizing and standardizing data to ensure consistency. The project also includes examples of how to use SQL queries to merge and join data from multiple tables, as well as how to filter and select specific rows based on certain conditions.

Overall, the project provides a good example of how to use SQL queries to clean and manipulate data effectively. The techniques used in the project may be useful for anyone interested in data cleaning, data manipulation, or SQL programming.

Requirements

Overview

The project titled "House Price Prediction" is a Jupyter Notebook-based project that focuses on predicting house prices using various machine learning algorithms. It begins with importing and exploring a dataset of house prices, followed by data cleaning and pre-processing.

The author employs a wide range of machine learning algorithms to predict house prices, such as:

  • Linear regression
  • Decision trees
  • Random forests
  • Gradient boosting The author uses various techniques to evaluate the performance of the models, such as mean squared error (MSE) and root mean squared error (RMSE). Additionally, the author uses various data visualization techniques to help better understand the data and the model's predictions.

The project also includes examples of how to use feature selection techniques to select the most important features in the dataset, which can help improve the accuracy of the models.

Overall, the project provides a good example of how to use machine learning algorithms to predict house prices and may be useful for anyone interested in machine learning, data analysis, or real estate.

Requirements

Overview

The Amazon Web Scraper Project is a Python-based project that focuses on scraping data from Amazon's website using the BeautifulSoup library. It can automatically extract product data from Amazon 86,400 times a day through Python libraries and generate product Dataset The project involves collecting information on product names, prices, ratings, and descriptions from Amazon's Best Sellers page. The data collected is then stored in a CSV file for future analysis. The code is heavily commented, making it easy for beginners to understand, and provides ample opportunities for customization or further development. The project is an excellent way to learn about web scraping, data handling, and Python programming as a whole. By the end of the project, developers will have gained a valuable skillset that they can apply to numerous other projects.

Requirements

Overview

The project focuses on analyzing a dataset of movie ratings and reviews using Python. The project begins by importing necessary libraries such as pandas, matplotlib, and seaborn.

The author then proceeds to clean and pre-process the dataset to ensure that it is in a format suitable for analysis. The pre-processing techniques used include handling missing values, removing duplicate data, and transforming data types.

The author uses various data visualization techniques to help better understand the data, such as creating histograms, scatter plots, and heatmaps. The project also includes examples of how to use statistical analysis to gain insights from the data, such as calculating mean, median, and mode.

Overall, the project provides a good example of how to use data analysis gain insights from a dataset of movie ratings and reviews. The techniques used in the project may be useful for anyone interested in data analysis, or the movie industry.

Requirements

barkamol_portfolio's People

Contributors

barkamoljon avatar

Stargazers

 avatar

Watchers

 avatar

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    ๐Ÿ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. ๐Ÿ“Š๐Ÿ“ˆ๐ŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google โค๏ธ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.