NumPy
Pandas
| NumPy | Pandas | |
|---|---|---|
| 310 | 431 | |
| 31,038 | 47,376 | |
| 1.0% | 0.8% | |
| 10.0 | 9.9 | |
| 4 days ago | about 10 hours ago | |
| Python | Python | |
| GNU General Public License v3.0 or later | BSD 3-clause "New" or "Revised" License |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
NumPy
- Python is not a great language for data science. Part 1: The experience
- Choosing Tech Stack in 2025: A Practical Guide
Unmatched integration with ML/AI ecosystems through NumPy, TensorFlow, and PyTorch
- What Dynamic Typing Is For
- Bringing NumPy's type-completeness score to nearly 90% – Pyrefly
> Let’s take a pause here for a second - the ‘CanIndex’ and ‘SupportsIndex’ from the looks are just “int”.
The PR for the change is https://github.com/numpy/numpy/pull/28913 - The details of files changed[0] shows the change was made in 'numpy/__init__.pyi'. Looking at the whole file[1] shows SupportsIndex is being imported from the standard library's typing module[2].
Where are you seeing SupportsIndex being defined as an int?
> I have a hard time dealing with these custom types because they are so obscure.
SupportsIndex is obscure, I agree, but it's not a custom type. It's defined in stdlib's typing module[2], and was added in Python 3.8.
[0]: https://github.com/numpy/numpy/pull/28913/files
[1]: https://github.com/charris/numpy/blob/c906f847f8ebfe0adec896...
[2]: https://docs.python.org/3/library/typing.html#typing.Support...
- Don’t Let Cyber Risk Kill Your GenAI Vibe: A Developer’s Guide
Know (or check) tells of older versions, such as the python sdk of OpenAI changing from a client with global state in v0.x.x, to a declared instance in v1.x.x, or numpy's change in how random generators are declared.
- Top 5 GitHub Repositories for Data Science in 2026
The book introduces the core libraries essential for working with data in Python: particularly IPython, NumPy, Pandas, Matplotlib, Scikit-Learn, and related packages Familiarity with Python as a language is assumed; if you need a quick introduction to the language itself, see the free companion project, A…
- Your 2025 Roadmap to Becoming an AI Engineer for Free for Vue.js Developers
AI starts with math and coding. You don’t need a PhD—just high school math like algebra and some geometry. Linear algebra (think matrices) and calculus (like slopes) help understand how AI models work. Python is the main language for AI, thanks to tools like TensorFlow and NumPy. If you know JavaScript from Vue.js, Python’s syntax is straightforward.
- Top 17 Tools for Scientific Simulation & Modeling
- Release v2.3.0 (June 7, 2025) · NumPy/NumPy
- How to Get Started with Scikit-Learn: A Beginner-Friendly Guide to Machine Learning in Python
As is the case with most Python libraries, it is open-source and free-to-use, making it easily accessible by anyone willing to learn machine learning, and it is built upon other open-source libraries within Python, like SciPy for advanced scientific operations, NumPy for efficient numerical computations, Matplotlib for data visualization, and Cython for increased efficiency and speed, similar to that of C/C++.
Pandas
- How to Analyze CSV Files with Python and Pandas
That’s where Python and Pandas shine. Pandas is a Python library that makes it easy to load, clean, analyze, and visualize data.
- Node.js vs Python: Real Benchmarks, Performance Insights, and Scalability Analysis
data analytics stacks (Pandas)
- Top 5 GitHub Repositories for Data Science in 2026
The book introduces the core libraries essential for working with data in Python: particularly IPython, NumPy, Pandas, Matplotlib, Scikit-Learn, and related packages Familiarity with Python as a language is assumed; if you need a quick introduction to the language itself, see the free companion project, A…
- Writing memory efficient C structs
https://github.com/pandas-dev/pandas/issues/58062 :
> On disk Parquet appears to store the category data as logical type String which is compressed with snappy and encoded
Arrow Flight RPC handles nested structs with enums over the wire somehow too FWIU
- Don't Know These 6 Tools? No Wonder Your Python Development Is So Slow
👉 https://pandas.pydata.org/
- Open Source Can't Coordinate
- Top Programming Languages for AI Development in 2025
Libraries for data science and deep learning that are always changing
- How to import sample data into a Python notebook on watsonx.ai and other questions…
# Read the content of nda.txt try: import os, types import pandas as pd from botocore.client import Config import ibm_boto3 def __iter__(self): return 0 # @hidden_cell # The following code accesses a file in your IBM Cloud Object Storage. It includes your credentials. # You might want to remove those credentials before you share the notebook. cos_client = ibm_boto3.client(service_name='s3', ibm_api_key_id='api-generated', ibm_auth_endpoint="https://iam.cloud.ibm.com/identity/token", config=Config(signature_version='oauth'), endpoint_url='https://s3.direct.us-south.cloud-object-storage.appdomain.cloud') bucket = 'your-bucket-referenced-here' object_key = 'nda__da__crxq8b2hmy.txt' # load data of type "text/plain" into a botocore.response.StreamingBody object. # Please read the documentation of ibm_boto3 and pandas to learn more about the possibilities to load the data. # ibm_boto3 documentation: https://ibm.github.io/ibm-cos-sdk-python/ # pandas documentation: http://pandas.pydata.org/ streaming_body_1 = cos_client.get_object(Bucket=bucket, Key=object_key)['Body'] with open("nda.txt", "r") as f: nda_content = f.read() print("Content of nda.txt has been read.") except FileNotFoundError: print("Error: nda.txt not found in the current directory.") nda_content = "" # Initialize knowledge source content_source = CrewDoclingSource( file_paths=["..."] )
- How I Hacked Uber’s Hidden API to Download 4379 Rides
As with any web scraping or data processing project, I had to write a fair amount of code to clean this up and shape it into a format I needed for further analysis. I used a combination of Pandas and regular expressions to clean it up (full code here).
- Show HN: Aiopandas – Async .apply() and .map() for Pandas, Faster API/LLMs Calls
Can this be merged into pandas?
Pandas does not currently install tqdm by default.
pandas-dev/pandas//pyproject.toml [project.optional-dependencies] https://github.com/pandas-dev/pandas/blob/main/pyproject.tom...
Dask solves for various adjacent problems; IDK if pandas, dask, or dask-cudf would be faster with async?
Dask docs > Scheduling > Dask Distributed (local) https://docs.dask.org/en/stable/scheduling.html#dask-distrib... :
> Asynchronous Futures API
Dask docs > Deploy Dask Clusters; local multiprocessing poll, k8s (docker desktop, podman-desktop,), public and private clouds, dask-jobqueue (SLURM,), dask-mpi:
What are some alternatives?
mitmproxy - An interactive TLS-capable intercepting HTTP proxy for penetration testers and software developers.
Airflow - Apache Airflow - A platform to programmatically author, schedule, and monitor workflows
SymPy - A computer algebra system written in pure Python
Cubes - [NOT MAINTAINED] Light-weight Python OLAP framework for multi-dimensional data analysis
Numba - NumPy aware dynamic Python compiler using LLVM
orange - 🍊 :bar_chart: :bulb: Orange: Interactive data analysis