Building a reliable product spreadsheet for Chinese e-commerce platforms is not a simple scraping exercise. It requires iterative refinement based on real ordering experience, and honest acknowledgment of what can and cannot be verified at scale. Here's a transparent breakdown of the methodology, including the failures and compromises that shaped the final approach.
Product Selection Criteria
Products enter the spreadsheet through a multi-stage funnel. The initial pool comes from category searches on Taobao, 1688, and Weidian, filtered by minimum seller ratings (4.6+ on a 5-point scale) and minimum transaction counts (typically 100+ sales for the specific listing). This eliminates the vast majority of listings immediately—most products on Chinese platforms come from new or low-volume sellers.
From this filtered pool, products are manually reviewed for listing quality: clear product photos, complete specifications, reasonable pricing relative to category averages, and coherent descriptions. Products with obvious red flags—watermarked images from other sellers, prices significantly below manufacturing costs, or nonsensical machine-translated descriptions—are excluded at this stage.
The criteria deliberately favor established sellers over new ones, even though this means missing some legitimate budget options. The tradeoff is lower risk for users who rely on the spreadsheet for initial product discovery.
Rejection and Removal Rules
Products are removed from the spreadsheet under specific conditions: seller rating drops below threshold, multiple user reports of quality issues, listing becomes unavailable, or price increases beyond 50% of original listing. Removal is not instant—a single complaint doesn't trigger removal, but patterns do.
Products are also removed if test orders reveal systematic problems. For example, if a watch listing advertises "316L stainless steel" but test purchases consistently show corrosion within weeks, that product gets flagged and eventually removed regardless of seller rating.
Failures and Mistakes Made
The early versions of this spreadsheet methodology had significant problems. The first iteration relied too heavily on automated scraping without manual review, resulting in listings that looked legitimate but came from sellers who simply copied successful competitors' photos. Several test orders from that period arrived as completely different products than pictured.
Another early mistake was trusting platform ratings without context. A seller might have 4.9 stars on 10,000 reviews for shipping speed and communication, but their actual product quality reviews (buried in the detailed breakdown) might be significantly lower. The current methodology weights product-specific ratings more heavily than overall seller ratings.
Perhaps the most costly lesson was underestimating seasonal variance. Products that tested well during summer production runs showed quality drops during Chinese New Year periods when factories use temporary workers. This isn't something spreadsheet data can fully capture, but it's documented in the notes where relevant.
Tradeoffs and Limitations
There's an inherent tension between spreadsheet comprehensiveness and accuracy. A spreadsheet with 50,000 products cannot have the same verification depth as one with 500 products. The current approach prioritizes breadth for discovery purposes while concentrating verification efforts on high-interest categories and popular items.
Price accuracy is another compromise. Checking prices for thousands of products weekly is not feasible. Instead, prices are snapshot values that drift over time. Users should treat spreadsheet prices as ballpark figures, not quotes. The actual price might be 10-20% different by the time you order.
The spreadsheet also cannot capture intangible factors like seller responsiveness or dispute resolution quality. A seller might have great products but terrible communication when problems arise. This kind of information only emerges through actual ordering experience, which can't be systematized into spreadsheet columns.
Bias Control Decisions
Affiliate relationships create obvious bias potential. Products from platforms or sellers with referral arrangements could theoretically receive preferential placement. To control for this, the spreadsheet uses standardized inclusion criteria that don't reference commercial relationships. A product either meets the quality and seller thresholds or it doesn't—affiliate status doesn't factor into inclusion.
However, it would be naive to claim complete objectivity. Personal product preferences inevitably influence which categories receive more attention during the curation process. Fashion products get more granular categorization than industrial supplies simply because there's more user interest and testing depth in those areas.
What Was Tested and Discarded
Several data points were tested for inclusion but ultimately dropped. Estimated shipping times proved too variable to be useful—the same seller might ship in 2 days or 2 weeks depending on stock and season. User review summaries were tried but Chinese reviews are often incentivized or fake, making aggregation misleading.
Real-time stock indicators were technically possible through API scraping but created maintenance overhead that wasn't justified by accuracy improvements. Sellers frequently show products as "in stock" that are actually made-to-order, making the data unreliable anyway. The current approach acknowledges stock uncertainty rather than providing false precision.
Summary: Why Methodology Matters
The value of a product spreadsheet isn't just the data it contains—it's the thinking behind what data is included, how it's verified, and what limitations are acknowledged. A spreadsheet built on automated scraping alone will contain garbage. A spreadsheet with no commercial relationships might be more "pure" but also less maintained. Understanding these tradeoffs helps you evaluate whether any given spreadsheet deserves your trust.