cookie

Utilizamos cookies para mejorar tu experiencia de navegación. Al hacer clic en "Aceptar todo", aceptas el uso de cookies.

avatar

Data Scientists | Machine Learning | AI Chatbots | Prompt Engineering

Perfect channel to learn Data Analytics, Data Sciene, Machine Learning & Artificial Intelligence Buy ads: https://telega.io/c/datasciencefree

Mostrar más
Publicaciones publicitarias
30 390
Suscriptores
+14224 horas
+1 1847 días
+4 69930 días

Carga de datos en curso...

Tasa de crecimiento de suscriptores

Carga de datos en curso...

How is 𝗖𝗜/𝗖𝗗 𝗽𝗿𝗼𝗰𝗲𝘀𝘀 𝗱𝗶𝗳𝗳𝗲𝗿𝗲𝗻𝘁 𝗳𝗼𝗿 𝗠𝗮𝗰𝗵𝗶𝗻𝗲 𝗟𝗲𝗮𝗿𝗻𝗶𝗻𝗴 𝗣𝗿𝗼𝗷𝗲𝗰𝘁𝘀 compared to 𝗥𝗲𝗴𝘂𝗹𝗮𝗿 𝘀𝗼𝗳𝘁𝘄𝗮𝗿𝗲? The important difference that the Machine Learning aspect of the projects brings to the CI/CD process is the treatment of the Machine Learning Training pipeline as a first class citizen of the software world. ➡️ CI/CD pipeline is a separate entity from Machine Learning Training pipeline. There are frameworks and tools that provide capabilities specific to Machine Learning pipelining needs (e.g. KubeFlow Pipelines, Sagemaker Pipelines etc.). ➡️ ML Training pipeline is an artifact produced by Machine Learning project and should be treated in the CI/CD pipelines as such. What does it mean? Let’s take a closer look: Regular CI/CD pipelines will usually be composed of at-least three main steps. These are: 𝗦𝘁𝗲𝗽 𝟭: Unit Tests - you test your code so that the functions and methods produce desired results for a set of predefined inputs. 𝗦𝘁𝗲𝗽 𝟮: Integration Tests - you test specific pieces of the code for ability to integrate with systems outside the boundaries of your code (e.g. databases) and between the pieces of the code itself. 𝗦𝘁𝗲𝗽 𝟯: Delivery - you deliver the produced artifact to a pre-prod or prod environment depending on which stage of GitFlow you are in. What does it look like when ML Training pipelines are involved? 𝗦𝘁𝗲𝗽 𝟭: Unit Tests - in mature MLOps setup the steps in ML Training pipeline should be contained in their own environments and Unit Testable separately as these are just pieces of code composed of methods and functions. 𝗦𝘁𝗲𝗽 𝟮: Integration Tests - you test if ML Training pipeline can successfully integrate with outside systems, this includes connecting to a Feature Store and extracting data from it, ability to hand over the ML Model artifact to the Model Registry, ability to log metadata to ML Metadata Store etc. This CI/CD step also includes testing the integration between each of the Machine Learning Training pipeline steps, e.g. does it succeed in passing validation data from training step to evaluation step. 𝗦𝘁𝗲𝗽 𝟯: Delivery - the pipeline is delivered to a pre-prod or prod environment depending on which stage of GitFlow you are in. If it is a production environment, the pipeline is ready to be used for Continuous Training. You can trigger the training or retraining of your ML Model ad-hoc, periodically or if the deployed model starts showing signs of Feature/Concept Drift.
Mostrar todo...
👍 4
Top Platforms for Building Data Science Portfolio Build an irresistible portfolio that hooks recruiters with these free platforms. Landing a job as a data scientist begins with building your portfolio with a comprehensive list of all your projects. To help you get started with building your portfolio, here is the list of top data science platforms. Remember the stronger your portfolio, the better chances you have of landing your dream job. 1. GitHub 2. Kaggle 3. LinkedIn 4. Medium 5. MachineHack 6. DagsHub 7. HuggingFace 7 Websites to Learn Data Science for FREE🧑‍💻 ✅ w3school ✅ datasimplifier ✅ hackerrank ✅ kaggle ✅ geeksforgeeks ✅ leetcode ✅ freecodecamp
Mostrar todo...
👍 19
Visualization.pdf5.29 MB
👍 12
$PONKE is one of the top memecoins on the solana blockchain Similar to $PEPE on ETH.  The Market cap of $PONKE is currently 240M while $PEPE is sitting at 6BILLION.  Very intresting token i've done some research and think this is one of the best bets in crypto to hold for long term. I started a positon at this level of  $0.44 cents and will hold until $1.00 is reached. The community is crazy and this reminds me of $SHIBA and $DOGE community  The social media accounts for PONKE are crazy  Instagram : https://tglink.io/de1108d4946f Twitter : https://tglink.io/b848e8b12c37 Telegram : https://tglink.io/96b36e41df1d Website : https://tglink.io/5e56265065fa https://tglink.io/85cba56cb05a https://tglink.io/356ed47406aa
Mostrar todo...
👍 4
SQL Quick Revision Notes 👇👇 https://topmate.io/analyst/864817
Mostrar todo...
SQL Quick Revision Notes for Interviews with Data Analyst

Data Analyst | with 30000+ followers | I guide people aspiring for data analytics and BI roles | Data Science

How to enter into Data Science 👉Start with the basics: Learn programming languages like Python and R to master data analysis and machine learning techniques. Familiarize yourself with tools such as TensorFlow, sci-kit-learn, and Tableau to build a strong foundation. 👉Choose your target field: From healthcare to finance, marketing, and more, data scientists play a pivotal role in extracting valuable insights from data. You should choose which field you want to become a data scientist in and start learning more about it. 👉Build a portfolio: Start building small projects and add them to your portfolio. This will help you build credibility and showcase your skills.
Mostrar todo...
👍 26
Machine Learning 💡.pdf1.58 MB
👍 9
Essential Data Science Key Concepts 1. Data: Data is the raw information that is collected and stored. It can be structured (in databases or spreadsheets) or unstructured (text, images, videos). Data can be quantitative (numbers) or qualitative (descriptions). 2. Data Cleaning: Data cleaning involves identifying and correcting errors in the dataset, handling missing values, removing outliers, and ensuring data quality before analysis. 3. Data Exploration: Data exploration involves summarizing the main characteristics of the data, understanding data distributions, identifying patterns, and detecting correlations or relationships within the data. 4. Descriptive Statistics: Descriptive statistics are used to describe and summarize the main features of a dataset. This includes measures like mean, median, mode, standard deviation, and visualization techniques. 5. Data Visualization: Data visualization is the graphical representation of data to help in understanding patterns, trends, and insights. Common visualization tools include bar charts, histograms, scatter plots, and heatmaps. 6. Statistical Inference: Statistical inference involves drawing conclusions from data with uncertainty. It includes hypothesis testing, confidence intervals, and regression analysis to make predictions or draw insights from data. 7. Machine Learning: Machine learning is a subset of artificial intelligence that uses algorithms to learn from data and make predictions or decisions without being explicitly programmed. It includes supervised learning, unsupervised learning, and reinforcement learning. 8. Feature Engineering: Feature engineering is the process of selecting, transforming, and creating features (input variables) to improve model performance in machine learning tasks. 9. Model Evaluation: Model evaluation involves assessing the performance of a machine learning model using metrics like accuracy, precision, recall, F1 score, ROC-AUC, and confusion matrix. 10. Data Preprocessing: Data preprocessing involves preparing the data for analysis or modeling. This includes encoding categorical variables, scaling numerical data, and splitting the data into training and testing sets.
Mostrar todo...