I built MockMyData.io in 10 weeks - a multi-tenant SaaS that lets developers generate mock REST APIs in seconds. Each project gets its own subdomain, API key, andI built MockMyData.io in 10 weeks - a multi-tenant SaaS that lets developers generate mock REST APIs in seconds. Each project gets its own subdomain, API key, and

I Built a Mock API Platform in 2.5 Months (Django + React + Redis + PostgreSQL)

The Problem

When building frontends, I didn't want to hardcode JSON data everywhere. I wanted to build against actual API endpoints and practice the integration work, but spinning up a real backend for every prototype felt like overkill.

After doing this dance too many times - writing throwaway Express servers or mocking fetch calls with static JSON - I realized other developers probably face this same friction.

The typical workflow looked like this:

1. Start building a React component

2. Realize I need user data

3. Choose between: hardcoded JSON arrays OR spinning up Express/Django

4. If I chose Django: pip install requirements, define models, run migrations, write views/URLs, configure CORS

5. All this before writing a single line of frontend logic

This context-switching killed momentum. I'd spend 30 minutes on backend setup for a feature that takes 10 minutes to prototype. Multiply this across dozens of projects, and you've lost days to infrastructure overhead.

That's why I built MockMyData.io.

What It Does

MockMyData.io lets developers generate mock REST API endpoints in seconds. When you sign up, you're automatically assigned a subdomain and API key. Then you define your data schema (field names and types) to create your endpoints.

You can either:

Enter custom JSON - Full control over your mock data

Auto-generate records - MockMyData.io creates realistic data based on your field types

Want to try it? Head to https://mockmydata.io - you can generate a demo mock API right from the landing page and start making requests immediately. No sign-up required.

Perfect for:

  • Frontend developers building features before backend APIs are ready
  • Mobile app developers testing API integration without backend dependencies
  • Building portfolio projects and demos without complex backend setup
  • Students and bootcamp grads learning API consumption
  • Rapid prototyping when you need to validate ideas fast

The Journey: 2.5 Months Solo

I went from concept to launch in about 10 weeks, building everything myself. The aggressive timeline kept me focused on shipping rather than over-engineering.

Tech Stack:

  • Backend: Django REST Framework
  • Frontend: React with Material-UI
  • Database: PostgreSQL
  • Caching: Redis
  • Payments: Stripe
  • Auth: Firebase SSO
  • Hosting: Render.com

Architecture Decisions

Multi-Tenant Subdomain Architecture

The core challenge was giving each project its own API endpoint while keeping infrastructure simple. I went with subdomain-based routing where users get automatically assigned subdomains on signup (e.g., `random-name-123.api.mockmydata.io`). Pro users can create custom subdomains and change them anytime.

How it works:

- User signs up and gets assigned a subdomain

- Pro users can customize: `mycompany.api.mockmydata.io`

- All API requests hit their subdomain endpoint

Implementation challenges:

1. DNS Management

Had to set up wildcard DNS records pointing all subdomains to the same server. Used Cloudflare for DNS with a wildcard A record (`*.api.mockmydata.io`).

2. SSL Certificates

Needed wildcard SSL certificates to handle unlimited subdomains. Render.com handles automatic SSL certificate provisioning and renewal for wildcard domains, which simplified deployment significantly.

3. Request Routing

Django's URL routing doesn't natively support subdomain-based tenant isolation. Built custom middleware to:

- Extract subdomain from request

- Look up project in database

- Attach project context to request object

- Route to appropriate data

# Simplified middleware example class SubdomainMiddleware: def init(self, get_response): self.get_response = get_response def call(self, request): subdomain = request.get_host().split('.')[0] try: project = Project.objects.get(subdomain=subdomain) request.project = project except Project.DoesNotExist: return HttpResponse('Project not found', status=404) return self.get_response(request)

Database Design

Used PostgreSQL with a shared schema approach rather than separate databases per tenant. Each endpoint and record has a `project_id` foreign key. This keeps infrastructure simple while maintaining data isolation through application-level filtering.

Why not separate databases per tenant?

- Simpler infrastructure (one database to manage)

- Easier backups and migrations

- Cost-effective for free tier users

- Row-level security handles isolation

Technical Challenge

#1: Multi-Tier Rate Limiting

This was trickier than expected. I needed two types of rate limiting:

Daily Quotas - Tier-based limits

- Free: 100 requests/day,

- Pro: Unlimited

Request Throttling - Spam prevention

- All tiers: Max 60 requests/minute

Why Redis?

Needed atomic increments and TTL support. Redis handles both perfectly:

def check_daily_quota(self, user): """Hard limits for FREE users""" daily_count = self.get_daily_count(user.id) DAILY_LIMIT = 100 if daily_count >= DAILY_LIMIT: return False, f'Daily limit reached ({DAILY_LIMIT} requests/day)' return True, 'OK' # CloudFlare helps with rate limiting and makes it simple and reliable

The Challenge: Making this performant at scale

- Redis calls add latency

- Need to fail fast for rate-limited requests

- Must be accurate (can't lose count data)

Solution: Batched Redis commands using pipelines reduced roundtrips, cutting rate-check latency significantly. I also implemented a circuit breaker pattern - if Redis goes down, requests pass through to prevent complete service outage.

Technical Challenge #2: Handling Pro-to-Free Downgrades

Free users can create up to 3 endpoints. However, when Pro users downgrade to Free, they might already have dozens of endpoints created. Rather than force them to delete endpoints, I let them choose which 3 remain active and accessible via API.

This required:

- Real-time enforcement in middleware before database queries

- Caching to avoid N+1 queries on every API request

- Graceful Redis fallback if caching fails

The system checks endpoint access on every API request:

#(Sample code) def _check_endpoint_selection(self, request, user, project):     """Check if endpoint is accessible for downgraded free users"""     # Pro users: all endpoints accessible     if user.is_pro_active:         return True, None       # Count total endpoints (cached)     endpoint_count = cache.get(f'project:{project.id}:endpoint_count')     # If <=3 endpoints total, all are accessible     if endpoint_count <= 3:         return True, None     # They have >3 endpoints (downgraded from Pro)     # Check if THIS endpoint is in their selected 3     endpoint = cache.get(f'endpoint:{project.id}:{path}')     if not endpoint.default_selected:         return False, JsonResponse({             'error': 'Endpoint Not Selected',             'message': 'This endpoint is not in your active selection. Free users can only have 3 active endpoints.',             'action_required': 'Visit your dashboard to manage active endpoints'         }, status=403)  # If cache miss then we fetch from database

This gracefully handles downgrades without data loss - users keep all their endpoints but must choose which 3 are live.

Technical Challenge #3: Handling Anonymous Demo Endpoints

Users can create temporary mock APIs without signing up. These urls expire within a short time and have strict limits on the total requests. The challenge was:

- Storing temporary projects in Redis (not database)

- Enforcing limits without database writes

- Supporting full CRUD operations on anonymous data

- Updating Redis cache after POST/PUT/PATCH/DELETE

All anonymous endpoints get a `demo-` prefix and live entirely in Redis with proper cache updates after mutations.

Technical Challenge #4: Storage Limits & Payload Validation

Implemented 4-layer protection to prevent abuse:

Layer 1: Request payload size

- Free: 5KB per request

- Pro: 30KB per request

Layer 2: Individual field size

- Free: 2KB per field

- Pro: 10KB per field

Layer 3: Item count per endpoint

- Free: 20 items

- Pro: 200 items

Layer 4: Total endpoint storage

- Free: 15KB per endpoint

- Pro: 400KB per endpoint

This prevents users from storing massive datasets while keeping the service performant and cost-effective.

What's Next: Django Project Generator

I'm building a feature that exports your mock API as production-ready backend code starting with Django. Here's how it works:

Input: Your MockMyData.io project with endpoints defined

Output: Complete Django REST Framework project with:

- Models generated from your schema

- Serializers for each endpoint

- CRUD operations

- URL routing configured

- Authentication setup (Optional)

- README with additional instructions & Suggestions

Example transformation:

Your MockMyData endpoint:

{ "name": "users", "fields": { "username": "string", "email": "email", "age": "integer" } }

Generates Django model:

class User(models.Model): username = models.CharField(max_length=255) email = models.EmailField() age = models.IntegerField() created_at = models.DateTimeField(auto_now_add=True)

Plus serializers, views, and URLs - everything a user need to run their server right away. I also plan on exploring other backends too

Why this matters:

Turns MockMyData.io from a testing tool into a full development accelerator. Prototype with mock data, export to production code when ready.

Lessons Learned

1. Ship fast, iterate faster

The 2.5-month timeline was aggressive but kept me focused on shipping. Rather than building every possible feature upfront, I launched with the core product working and plan to aggressively iterate based on what users actually need.

2. Rate limiting is harder than you think

Especially across multiple tiers and preventing race conditions. Redis pipelines were essential.

3. Cache everything strategically

Redis saved my infrastructure costs. Without caching, I'd be paying 3-4x more for database and compute.

4. Stripe webhooks are your friend

Once you understand them. The documentation is excellent, and webhook-driven subscription management is reliable

6. Build for failure

My circuit breaker pattern for Redis means the service stays up even when caching fails. Graceful degradation is better than complete outages.

Try It Out

🚀 [https://mockmydata.io]() - Free tier available, no credit card required

🎉 Launching on Product Hunt January 14th - Would love your support!

💬 Questions I'd love feedback on:

What backend frameworks would you want for code export? (Django, Express, FastAPI, Rails?)

What's missing that would make this a must-have tool for you?

Drop a comment below - happy to answer questions about Django, React, multi-tenant architecture, or building a SaaS solo! You can also connect with me @marcuscodes.

Market Opportunity
Wrapped REACT Logo
Wrapped REACT Price(REACT)
$0.04683
$0.04683$0.04683
-1.41%
USD
Wrapped REACT (REACT) Live Price Chart
Disclaimer: The articles reposted on this site are sourced from public platforms and are provided for informational purposes only. They do not necessarily reflect the views of MEXC. All rights remain with the original authors. If you believe any content infringes on third-party rights, please contact service@support.mexc.com for removal. MEXC makes no guarantees regarding the accuracy, completeness, or timeliness of the content and is not responsible for any actions taken based on the information provided. The content does not constitute financial, legal, or other professional advice, nor should it be considered a recommendation or endorsement by MEXC.

You May Also Like

Is Doge Losing Steam As Traders Choose Pepeto For The Best Crypto Investment?

Is Doge Losing Steam As Traders Choose Pepeto For The Best Crypto Investment?

The post Is Doge Losing Steam As Traders Choose Pepeto For The Best Crypto Investment? appeared on BitcoinEthereumNews.com. Crypto News 17 September 2025 | 17:39 Is dogecoin really fading? As traders hunt the best crypto to buy now and weigh 2025 picks, Dogecoin (DOGE) still owns the meme coin spotlight, yet upside looks capped, today’s Dogecoin price prediction says as much. Attention is shifting to projects that blend culture with real on-chain tools. Buyers searching “best crypto to buy now” want shipped products, audits, and transparent tokenomics. That frames the true matchup: dogecoin vs. Pepeto. Enter Pepeto (PEPETO), an Ethereum-based memecoin with working rails: PepetoSwap, a zero-fee DEX, plus Pepeto Bridge for smooth cross-chain moves. By fusing story with tools people can use now, and speaking directly to crypto presale 2025 demand, Pepeto puts utility, clarity, and distribution in front. In a market where legacy meme coin leaders risk drifting on sentiment, Pepeto’s execution gives it a real seat in the “best crypto to buy now” debate. First, a quick look at why dogecoin may be losing altitude. Dogecoin Price Prediction: Is Doge Really Fading? Remember when dogecoin made crypto feel simple? In 2013, DOGE turned a meme into money and a loose forum into a movement. A decade on, the nonstop momentum has cooled; the backdrop is different, and the market is far more selective. With DOGE circling ~$0.268, the tape reads bearish-to-neutral for the next few weeks: hold the $0.26 shelf on daily closes and expect choppy range-trading toward $0.29–$0.30 where rallies keep stalling; lose $0.26 decisively and momentum often bleeds into $0.245 with risk of a deeper probe toward $0.22–$0.21; reclaim $0.30 on a clean daily close and the downside bias is likely neutralized, opening room for a squeeze into the low-$0.30s. Source: CoinMarketcap / TradingView Beyond the dogecoin price prediction, DOGE still centers on payments and lacks native smart contracts; ZK-proof verification is proposed,…
Share
BitcoinEthereumNews2025/09/18 00:14
Fed Decides On Interest Rates Today—Here’s What To Watch For

Fed Decides On Interest Rates Today—Here’s What To Watch For

The post Fed Decides On Interest Rates Today—Here’s What To Watch For appeared on BitcoinEthereumNews.com. Topline The Federal Reserve on Wednesday will conclude a two-day policymaking meeting and release a decision on whether to lower interest rates—following months of pressure and criticism from President Donald Trump—and potentially signal whether additional cuts are on the way. President Donald Trump has urged the central bank to “CUT INTEREST RATES, NOW, AND BIGGER” than they might plan to. Getty Images Key Facts The central bank is poised to cut interest rates by at least a quarter-point, down from the 4.25% to 4.5% range where they have been held since December to between 4% and 4.25%, as Wall Street has placed 100% odds of a rate cut, according to CME’s FedWatch, with higher odds (94%) on a quarter-point cut than a half-point (6%) reduction. Fed governors Christopher Waller and Michelle Bowman, both Trump appointees, voted in July for a quarter-point reduction to rates, and they may dissent again in favor of a large cut alongside Stephen Miran, Trump’s Council of Economic Advisers’ chair, who was sworn in at the meeting’s start on Tuesday. It’s unclear whether other policymakers, including Kansas City Fed President Jeffrey Schmid and St. Louis Fed President Alberto Musalem, will favor larger cuts or opt for no reduction. Fed Chair Jerome Powell said in his Jackson Hole, Wyoming, address last month the central bank would likely consider a looser monetary policy, noting the “shifting balance of risks” on the U.S. economy “may warrant adjusting our policy stance.” David Mericle, an economist for Goldman Sachs, wrote in a note the “key question” for the Fed’s meeting is whether policymakers signal “this is likely the first in a series of consecutive cuts” as the central bank is anticipated to “acknowledge the softening in the labor market,” though they may not “nod to an October cut.” Mericle said he…
Share
BitcoinEthereumNews2025/09/18 00:23
Stronger capital, bigger loans: Africa’s banking outlook for 2026

Stronger capital, bigger loans: Africa’s banking outlook for 2026

African banks spent 2025 consolidating, shoring up capital, tightening risk controls, and investing in digital infrastructure, following years of macroeconomic
Share
Techcabal2026/01/14 23:06