Skip to content

Commit 89f6173

Browse files
authored
Merge pull request #32 from karaposu/dev
v2.3.0: codebase restructure, test suite rewrite, new scrapers
2 parents 972091f + 447c0a0 commit 89f6173

393 files changed

Lines changed: 7755 additions & 15355 deletions

File tree

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

.github/workflows/publish.yml

Lines changed: 29 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,29 @@
1+
name: Publish to PyPI
2+
3+
on:
4+
release:
5+
types: [published]
6+
7+
jobs:
8+
publish:
9+
runs-on: ubuntu-latest
10+
steps:
11+
- uses: actions/checkout@v4
12+
- name: Set up Python
13+
uses: actions/setup-python@v5
14+
with:
15+
python-version: "3.9"
16+
17+
- name: Install build dependencies
18+
run: |
19+
python -m pip install --upgrade pip
20+
pip install build twine
21+
22+
- name: Build package
23+
run: python -m build
24+
25+
- name: Publish to PyPI
26+
env:
27+
TWINE_USERNAME: __token__
28+
TWINE_PASSWORD: ${{ secrets.PYPI_API_TOKEN }}
29+
run: twine upload dist/*

.gitignore

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -11,7 +11,7 @@ __pycache__/
1111
*.so
1212

1313

14-
.devodcs
14+
.devdocs
1515

1616
# Distribution / packaging
1717
.Python

CHANGELOG.md

Lines changed: 10 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,15 @@
11
# Bright Data Python SDK Changelog
22

3+
## Version 2.3.0 - Browser API, Scraper Studio, 175 Datasets
4+
5+
- **Browser API**: Connect to cloud Chrome via CDP WebSocket. SDK builds the `wss://` URL, you connect with Playwright/Puppeteer (`client.browser.get_connect_url()`)
6+
- **Scraper Studio**: Trigger and fetch results from custom scrapers built in Bright Data's IDE (`client.scraper_studio.run()`)
7+
- **75 more datasets**: Agoda, AutoZone, BBC, Best Buy, Bluesky, Booking, Costco, eBay, Etsy, GitHub, Google News/Play/Shopping, Home Depot, Kroger, Lowe's, Macy's, Microcenter, Ozon, Quora, Realtor, Reddit, Snapchat, TikTok Shop, Tokopedia, Vimeo, Wayfair, Wikipedia, Wildberries, X/Twitter, Yahoo Finance, Zoopla, and more — **175 total**
8+
- **Codebase cleanup**: Removed dead code and legacy abstractions — collapsed `datasets/client.py` from 1635 to 285 lines, fixed `ScrapeJob.to_result()` crash bug, cleaned up unused protocols, redundant config layers, and stale API modules
9+
- **Test suite rewrite**: Rebuilt test suite from scratch with 365 unit tests, shared fixtures via `conftest.py`, behavioral coverage focus — key modules now at 87–98% coverage (client, scrapers, SERP, sync client, job lifecycle)
10+
11+
---
12+
313
## Version 2.2.1 - 100 Datasets API
414

515
### ✨ New Features

MANIFEST.in

Lines changed: 0 additions & 6 deletions
This file was deleted.

benchmarks/bench_async_vs_sync.py

Lines changed: 0 additions & 1 deletion
This file was deleted.

benchmarks/bench_batch_operations.py

Lines changed: 0 additions & 1 deletion
This file was deleted.

benchmarks/bench_memory_usage.py

Lines changed: 0 additions & 1 deletion
This file was deleted.

docs/api-reference/.gitkeep

Whitespace-only changes.

0 commit comments

Comments
 (0)