- Published on
AI Era Survival Guide Part 5: The Future of Data Scientists - Crisis or Opportunity?
- Authors

- Name
- Youngju Kim
- @fjvbn20031
AI Era Survival Guide Part 5: The Future of Data Scientists - Crisis or Opportunity?
In 2019, a Harvard Business Review article declaring "data scientist is the sexiest job of the 21st century" captured the imagination of IT professionals worldwide. Many people went back to study statistics, learned Python, and pulled all-nighters on Kaggle. Now, in 2026, LLMs like ChatGPT and Claude can produce data analysis reports in seconds, and AutoML tools automate model selection and hyperparameter tuning.
Let me ask you directly: "So what should data scientists do now?"
This article is for those who can't stop asking that question. The short answer is: it's both a crisis and an enormous opportunity. But which one it becomes depends entirely on the choices you make.
1. The Reality: What AutoML and LLMs Are Replacing
First, we need to face some uncomfortable truths. Let's take an honest look at the data scientist work that AI is already replacing.
Automated Exploratory Data Analysis (EDA)
Until 2023, EDA was a core task for data scientists: loading data, checking for missing values, visualizing distributions, and analyzing correlations. What about now?
# The old way: EDA written manually in dozens of lines
import pandas as pd
import matplotlib.pyplot as plt
import seaborn as sns
df = pd.read_csv('data.csv')
print(df.describe())
print(df.isnull().sum())
df.hist(figsize=(20, 15))
plt.show()
# ... dozens more lines
Today, you can ask Claude or GPT-4 to "analyze this dataset" and receive completed EDA code along with insights. Tools like ydata-profiling (formerly pandas-profiling) generate hundreds of pages of EDA reports from a single line of code.
The Democratization of Basic Model Building
The progress of AutoML is even more striking. Google's Vertex AI AutoML, Amazon SageMaker Autopilot, and open-source tools like Auto-sklearn and FLAML now allow even non-specialists to build reasonably good predictive models.
# Training a model with AutoML - just feed in the data
from flaml import AutoML
automl = AutoML()
automl.fit(
X_train,
y_train,
task="classification",
time_budget=3600 # Search for the best model within 1 hour
)
print(automl.best_estimator)
print(automl.best_config)
This used to be a task that required a junior DS several weeks to complete. Now it's done with one click.
Automated Routine Report Writing
Monthly performance reports that were once written in the same format every time are now automatically generated by LLMs from raw data. Many companies have already built automated reporting pipelines, and in the process, DS positions dedicated to mechanically writing reports are shrinking.
The Shift in Feature Engineering
Advances in deep learning have made automated feature learning possible even for tabular data. Models like TabNet and FT-Transformer achieve excellent performance without manual feature engineering. LLM-based feature engineering tools automatically translate features described in natural language into code.
Conclusion: A significant portion of routine and repetitive DS work is being automated. This is reality.
2. Core DS Competencies That Are Still Essential
But here is where an important turning point emerges. There are domains that automation cannot replace, and the value of those domains is actually increasing.
Domain Understanding and Business Problem Definition
AutoML can train on data. But defining "what problem should we solve?" remains a human domain.
Consider a healthcare example. When the business goal is "reduce patient readmission rates," translating that into a predictive modeling problem requires countless decisions:
- Whether to define readmission as within 30 days or 90 days
- Whether to target only preventable readmissions
- When to set the prediction point based on what interventions are possible
- Whether false positives or false negatives carry a higher cost
These decisions cannot be made without deep domain knowledge of medicine. And these decisions determine the actual business value of the model.
Experimental Design and Causal Inference
"Revenue went up. Was it because of the A/B test, or seasonal factors?" Answering that question is causal inference.
The ability to distinguish correlation from causation is an area that AI still struggles to replace. No matter how good a model you build from data collected through a poorly designed experiment, the conclusions will be meaningless.
# Causal inference: estimating treatment effects (illustrative concept of IV method)
# An approach for estimating the X -> Y causal relationship without confounders
# Simple OLS can be biased due to confounders
# Causal estimation using natural experiments or instrumental variables is important
from sklearn.linear_model import LinearRegression
import numpy as np
# Rather than a simple comparison of treatment and control groups,
# estimation corrected for selection bias is required.
# Making this judgment is a collaboration between domain experts and DS.
Methods like Difference-in-Differences, Synthetic Control, and Regression Discontinuity Design are steadily growing in value in business settings.
The Ability to Evaluate Data Quality
"Can I trust this data?" is a question AI finds very difficult to answer on its own. Understanding the context in which data was collected, patterns of measurement error, and the possibility of selection bias is the domain of an experienced DS.
In practice, many ML projects fail not because of algorithmic issues but because of data quality problems. No matter how good AutoML gets, the principle of "Garbage In, Garbage Out" doesn't change.
Quantifying and Communicating Uncertainty
Business decision-makers want to know "how certain is this model?" The ability to understand confidence intervals, prediction intervals, and model uncertainty, and to communicate them clearly to non-technical stakeholders, is still invaluable.
3. Evolving from DS to ML Engineer / AI Product Builder
Many DSs are choosing one of two directions.
Direction 1: Transitioning to ML Engineer / MLOps
The focus shifts from building models to operating models in production.
# Example of model tracking and deployment with MLflow
# mlproject file structure
name: churn_prediction
conda_env: conda.yaml
entry_points:
train:
parameters:
learning_rate: { type: float, default: 0.01 }
n_estimators: { type: int, default: 100 }
command: 'python train.py --lr {learning_rate} --n_est {n_estimators}'
evaluate:
command: 'python evaluate.py'
An MLOps engineer is responsible for:
- Automating model training pipelines
- Building and managing feature stores
- Model serving infrastructure (latency, throughput)
- Model drift monitoring and retraining triggers
- A/B testing infrastructure
Demand for MLOps engineers in Korea has surged since 2024. Not just major tech companies like Naver, Kakao, and Coupang, but also fintech and healthcare startups are actively hiring for this role.
Direction 2: Evolving into an AI Product Builder
This is a role that has newly emerged in the LLM era. It combines the analytical ability of a DS with engineering skills to build AI products that actually work.
# Example of a core RAG-based AI product pipeline
from langchain.embeddings import OpenAIEmbeddings
from langchain.vectorstores import Chroma
from langchain.chat_models import ChatOpenAI
from langchain.chains import RetrievalQA
# Document embedding and vector store setup
embeddings = OpenAIEmbeddings()
vectorstore = Chroma.from_documents(documents, embeddings)
# Retrieval-based QA chain
retriever = vectorstore.as_retriever(search_kwargs={"k": 5})
qa_chain = RetrievalQA.from_chain_type(
llm=ChatOpenAI(model="gpt-4o"),
chain_type="stuff",
retriever=retriever,
return_source_documents=True
)
# Execute query
result = qa_chain({"query": "What is the refund policy for our product?"})
This role goes beyond simple ML model development to designing and implementing entire AI systems.
4. AI Skills Data Scientists Should Learn
Understanding and Practicing LLM Fine-Tuning
Fine-tuning LLMs has become an important competency for DSs. Full fine-tuning is expensive, but using LoRA (Low-Rank Adaptation) makes it feasible even with consumer-grade GPUs.
# Example of LLM fine-tuning with LoRA (Hugging Face PEFT)
from peft import get_peft_model, LoraConfig, TaskType
from transformers import AutoModelForCausalLM, AutoTokenizer
model_name = "meta-llama/Llama-3.2-3B"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(model_name)
# LoRA configuration
lora_config = LoraConfig(
task_type=TaskType.CAUSAL_LM,
r=16, # LoRA rank
lora_alpha=32, # Scaling parameter
target_modules=["q_proj", "v_proj"],
lora_dropout=0.05,
bias="none",
)
# Apply LoRA adapter
peft_model = get_peft_model(model, lora_config)
peft_model.print_trainable_parameters()
# trainable params: 2,097,152 || all params: 3,213,000,000 || trainable%: 0.065
Knowing when fine-tuning is necessary versus when prompt engineering or RAG is sufficient is also an important skill.
RAG System Design and Implementation
Retrieval-Augmented Generation is currently the most widely used AI product pattern. A DS can take ownership of the following in a RAG system:
- Experimenting with chunking strategies (fixed-size vs. sentence-level vs. semantic-level)
- Selecting and evaluating embedding models
- Implementing hybrid search (vector search + BM25)
- Utilizing reranking models
# Hybrid search implementation concept
from langchain.retrievers import EnsembleRetriever
from langchain.retrievers import BM25Retriever
from langchain.vectorstores import FAISS
# Vector retriever
vector_retriever = faiss_store.as_retriever(search_kwargs={"k": 10})
# BM25 keyword retriever
bm25_retriever = BM25Retriever.from_documents(documents)
bm25_retriever.k = 10
# Ensemble retrieval (0.6: vector, 0.4: BM25)
ensemble_retriever = EnsembleRetriever(
retrievers=[vector_retriever, bm25_retriever],
weights=[0.6, 0.4]
)
AI System Evaluation
This is perhaps the most important new competency for DSs. LLM-based systems must be evaluated differently from traditional ML models. Fixed test-set accuracy is no longer sufficient.
# Example of an LLM system evaluation framework (RAGAS)
from ragas import evaluate
from ragas.metrics import (
faithfulness, # Is the generated answer faithful to the retrieved context?
answer_relevancy, # Is the answer relevant to the question?
context_precision, # Is the retrieved context accurate?
context_recall, # Were all relevant contexts retrieved?
)
result = evaluate(
dataset=test_dataset,
metrics=[faithfulness, answer_relevancy, context_precision, context_recall],
)
print(result.to_pandas())
Building evaluation datasets, combining human and automated evaluation, and constructing evaluation pipelines for regression prevention are all within the DS's domain.
Prompt Engineering and LLM Orchestration
Beyond simple prompt writing, you need the ability to design complex AI workflows.
# Agent workflow using LangGraph
from langgraph.graph import StateGraph, END
from typing import TypedDict, Annotated
import operator
class AgentState(TypedDict):
messages: Annotated[list, operator.add]
next: str
# Example analytical agent workflow
workflow = StateGraph(AgentState)
workflow.add_node("data_loader", load_data_node)
workflow.add_node("analyzer", analyze_node)
workflow.add_node("visualizer", visualize_node)
workflow.add_node("reporter", report_node)
workflow.set_entry_point("data_loader")
workflow.add_edge("data_loader", "analyzer")
workflow.add_conditional_edges(
"analyzer",
route_next_step,
{"needs_viz": "visualizer", "done": "reporter"}
)
workflow.add_edge("visualizer", "reporter")
workflow.add_edge("reporter", END)
5. A 12-Month Transition Roadmap
Let's build a concrete action plan.
Months 1-3: Assessing Your Foundation and Getting Started with LLMs
Month 1: Diagnosing Your Current State
- Categorize your current DS work into "at risk of automation" and "not at risk"
- Hands-on practice with LLM APIs (OpenAI API, Anthropic API)
- Read the Prompt Engineering Guide (https://www.promptingguide.ai) from cover to cover
Month 2: Understanding LLM Applications
- Complete a LangChain or LlamaIndex tutorial
- Build a simple RAG system yourself (practice feeding your domain knowledge to an AI)
- Take the free Hugging Face "LLM Course"
Month 3: Deciding on a Specialization Direction
- Choose between the MLOps direction and the AI product builder direction
- Begin focused study of the core tools for your chosen path
- Join communities: Korean MLOps community, LangChain Korea user group
Months 4-6: Acquiring Core Technical Skills
If you chose the MLOps path:
# Recommended learning sequence
# 1. Docker & Kubernetes basics
docker build -t my-model:v1 .
kubectl apply -f deployment.yaml
# 2. Master MLflow
mlflow server --backend-store-uri sqlite:///mlflow.db
# 3. Kubeflow or Vertex AI Pipelines
# 4. Monitoring: Prometheus + Grafana + Evidently AI
If you chose the AI product builder path:
- Deep dive into LangChain/LlamaIndex
- Understand vector databases (Pinecone, Weaviate, Chroma)
- LLM evaluation frameworks (RAGAS, DeepEval)
- Production deployment experience (Streamlit, FastAPI + Vercel)
Months 7-9: Real-World Projects
During this period, you must build something that actually works.
Recommended project ideas:
- A RAG chatbot specialized for your domain (healthcare, finance, e-commerce, etc.)
- A data analysis automation agent (automatically generates insights from CSV uploads)
- An MLOps pipeline project (GitHub Actions + MLflow + FastAPI)
Months 10-12: Positioning and Job Search
- Organize your projects on GitHub with thorough READMEs
- Post about your learning process and project retrospectives on a tech blog
- Update your LinkedIn profile
- Write a list of target companies and start applying
6. The Job Market Reality: Korea and Abroad
The Korean Market
Let's take an honest look at DS/AI hiring trends in Korea as of 2026.
Roles in high demand:
- MLOps/ML Engineer: Naver, Kakao, Coupang, Toss, Karrot Market, and more are all hiring
- LLM Application Engineer: Explosive demand, especially among startups
- AI Researcher (the kind who writes papers): Top university graduate school credentials and top-conference publications are required
Realistic barriers to entry:
- Large corporations still tend to prefer candidates from SKY/KAIST/POSTECH
- However, startups are increasingly hiring based on skill
- Portfolio and GitHub contributions are becoming more and more important
Salary reality (estimated for 2026):
- Junior DS/ML Engineer (1-3 years): 45M-65M KRW
- Mid-level (3-7 years): 70M-100M KRW
- Senior (7+ years): 100M+ KRW
- Significant upward adjustment possible with overseas experience or top-conference publications
The International Market (Including Remote Work)
If your English is solid, the international market is much wider.
- On platforms like Upwork and Toptal, you can work as an LLM consultant for $80-200/hour
- At remote-work startups, ML Engineer roles can pay 200K/year
- Kaggle Master/Grandmaster status is still a powerful signal abroad
7. Portfolio Strategy: From Kaggle to Real Products
Many DSs think Kaggle medals are the entirety of their portfolio. But the job market of 2026 has changed.
The Value and Limits of Kaggle
Kaggle is still valuable. Top placements prove your ML technical skills. But Kaggle problems are different from reality.
- Data is provided clean (reality is messy)
- Problems are clearly defined (in reality, defining the problem is step one)
- No need to think about deployment and operations (in reality, that's more than half the work)
Building a Real Product Portfolio
Level 1: Demonstrating Tool Mastery
Build and deploy a RAG system using data from your own field. For example:
- "Korean Drug Information Search AI" - using open data from the Ministry of Food and Drug Safety
- "Legal Document Summarization AI" - using publicly available case law
- "Financial Report Analysis AI" - using publicly disclosed financial data
# Deployed RAG system architecture
# 1. Data collection and preprocessing pipeline
# 2. Document chunking and embedding
# 3. Vector store construction
# 4. Retrieval + generation chain
# 5. FastAPI backend
# 6. Streamlit or React frontend
# 7. Docker containerization and cloud deployment (AWS/GCP/Azure)
Level 2: A Product with Real Users
Even 100 real users makes a product far more compelling.
- You get real usage data and feedback
- You gain experience solving problems that arise during operations
- It proves that "this person can actually build a product"
Level 3: Measurable Business Impact
The most powerful portfolio is one with concrete impact: "the AI I built reduced costs by X%" or "automated Y tasks." Try a small AI automation project at your current job. If it's hard to get permission, volunteering to help a small local business or nonprofit adopt AI is also an excellent portfolio piece.
Closing Thoughts
The job title of "data scientist" may disappear. But the demand for "people who understand data, define problems, and solve them with AI" will instead grow explosively.
If AutoML builds the model, we define which models should be built. If LLMs write the analysis report, we decide which questions to ask. If AI finds the patterns, we judge whether those patterns are meaningful.
It's okay if you don't have a perfect plan right now. Start by opening the OpenAI API documentation today. Build one small RAG project. That first step will completely transform your career one year from now.
The next installment in this series covers survival strategies for product managers in the AI era.