Discover more from Data Boutique
How Data Boutique Works
Making Web Data Purchases Easy
About Data Boutique
Data Boutique is a web-scraped data marketplace.
If you’re looking for web data, there is a high chance someone is already collecting it. Data Boutique makes it easier to buy web data from them.
Join our Platform to learn and interact about this project:
How Data Boutique Works
Data Boutique is a web data marketplace: Sellers offer data to Buyers, and the platform makes sure this exchange happens smoothly to both sides’ advantage.
Let’s see how data enters the store.
Call for Data
Data Boutique is meant to be a safe environment, where users can trust the data they find to be legitimately collected and suitable to their use. For this reason, unlike other marketplaces, where providers are free to upload any data, on Data Boutique data onboarding is subject to approval: Data needs to be “called”.
This means someone has filed a request to Data Boutique, and this request has been validated: A legitimate request is when it is legal to collect data from that source, and that source is in line with global standards of web scraping, being free from Personal Identifiable Information (PII) and Material-Nonpublic Information (MNPI).
Anyone can post a request:
Data Sellers: Those who might have the data, and wish to sell it on Data Boutique
Data Buyers: Those who might want the data, to use it in their applications.
Data Boutique evaluates all received requests. Once given a green light, it issues a Call for Data: A formal request for anyone who is technically and legally able to provide it.
Anyone can apply to a Call for Data:
Anyone who is doing web scraping
The website itself, when they want to monetize their data.
Anyone can apply… but only those who prove they can deliver will pass.
The admittance to the selection process follows a “first come first served” priority blended with the seller’s track record on Data Boutique.
When Data Boutique validates an application (the Seller is admitted to the selection process) the seller will automatically be provided with access keys and coordinates for that specific application.
The selection process requires a prompt, complete, and time-consistent flow of data. The onboarding is fully automated and validates the data structure, data completeness, and time consistency of daily data feeds.
Note: As part of our launch strategy, we are initially admitting a fixed quota of providers. Since we are receiving a large number of applications, we encourage future sellers for patience, everyone will be given a chance.
When proof of data delivery has been provided, the data feed is automatically set to LIVE and can be purchased at any moment by any Data Boutique buyer.
Data Boutique de-couples supply from demand. Data sellers provide daily refreshed files of their website content, regardless of what frequency buyers are getting the data.
This allows two main upsides:
Data is always updated, and the customer can purchase instantly the most recent content. Buyers can choose the frequency of the refresh of their purchases without the need to alert the seller, as the latter ones are serving at the highest frequency anyway;
The data content is checked daily against quality standards identical to those used during the selection process. Only files that pass the QA tests are pushed to the store.
The price of a single data purchase changes dynamically: It is higher for websites with fewer transactions - to support Sellers bearing the fixed cost of data extraction - and lower for websites with a high number of transactions per month.
This further incentivizes transactions on high-demand websites, accelerating the flywheel, where sellers’ gains are buyers’ savings, and where web scraping is more efficient, reducing its footprint on observed websites and energy consumption.
Join the Project
Data Boutique is a community for sustainable, ethical, high-quality web data exchanges. You can browse the current catalog and add your request if a website is not listed. Saving datasets to your interest list will allow sellers to correctly size the demand for datasets and onboard the platform.
More on this project can be found on our Discord channels.
Thanks for reading and sharing this.