Reddit Programming
210 subscribers
1.22K photos
124K links
I will send you newest post from subreddit /r/programming
Download Telegram
From Batch to Insights: How to Automate Data Validation Workflows
https://www.reddit.com/r/programming/comments/1np5m74/from_batch_to_insights_how_to_automate_data/

<!-- SC_OFF -->Hey r/programming (https://www.reddit.com/r/programming), I've been thinking a lot about the common pain points of dealing with unvalidated or "dirty" data, especially when working with large datasets. Manual cleaning is incredibly time-consuming and often a huge bottleneck for getting projects off the ground or maintaining data pipelines. It feels like a constant battle against inaccurate reports, compliance risks, and just generally wasted effort. Specifically, I'm looking into approaches for automating validation across different data types—like email addresses, mobile numbers, IP addresses, and even browser user-agents—for batch processing. Has anyone here implemented solutions using external APIs for this kind of batch data validation? What were your experiences? What are your thoughts on: * The challenges of integrating such third-party validation services? * Best practices for handling asynchronous batch processing (submission, polling, retrieval)? * The ROI you've seen from automating these processes versus maintaining manual checks or in-house solutions? * Any particular types of validation (e.g., email deliverability, mobile line type, IP threat detection) that have given you significant headaches or major wins with automation? Would love to hear about your experiences, cautionary tales, or success stories in building robust, automated data validation workflows! Cheers! <!-- SC_ON --> submitted by /u/Available-Floor9213 (https://www.reddit.com/user/Available-Floor9213)
[link] (https://www.onboardingbuddy.co/blog/from-batch-to-insights-data-validation) [comments] (https://www.reddit.com/r/programming/comments/1np5m74/from_batch_to_insights_how_to_automate_data/)
Scaling WhatsApp OTP delivery with Laravel + Redis (what we learned building CrunchzApp)
https://www.reddit.com/r/programming/comments/1npeazn/scaling_whatsapp_otp_delivery_with_laravel_redis/

<!-- SC_OFF -->Hey folks, Over the last few months I’ve been building CrunchzApp, a SaaS platform for sending WhatsApp OTPs and notifications at scale. Instead of pitching, I thought I’d share some of the technical challenges we ran into and how we solved them might be useful for others tackling queue-heavy or API-reliant systems. Stack: Laravel 12, InertiaJS React, MariaDB, Redis, Horizon. Challenges & solutions: Scaling message queues: OTPs need to be near-instant, but WhatsApp API calls can stall. We leaned on Redis + Horizon for distributed queues and optimized retry/backoff strategies. Channel load balancing: To avoid throttling, we built a round-robin algorithm that distributes messages across multiple WhatsApp channels. Testing safely: Every new channel automatically starts in a 7-day sandbox mode, tied to the subscription trial. This was tricky to design since it uses the same API surface as production, just with restrictions. Monitoring third-party reliability: WhatsApp sometimes delays or rejects messages. We had to build logging + alerting so developers can see exactly where the failure happens (our system, or WhatsApp). I’d love to get some discussion going on these points: If you’ve worked on queue-heavy apps, what’s your go-to approach for keeping jobs “real-time enough” under load? Any favorite strategies for monitoring external APIs when your SLA depends on them? How do you balance building developer-friendly APIs with maintaining internal complexity (sandboxing, routing, retries, etc.)? Curious to hear how others have approached similar problems 👀 <!-- SC_ON --> submitted by /u/masitings (https://www.reddit.com/user/masitings)
[link] (https://www.crunchz.app/) [comments] (https://www.reddit.com/r/programming/comments/1npeazn/scaling_whatsapp_otp_delivery_with_laravel_redis/)
Automating Vercel Deploys on Private Repos (Workaround Idea)
https://www.reddit.com/r/programming/comments/1npxaag/automating_vercel_deploys_on_private_repos/

<!-- SC_OFF -->So I’ve been playing with a problem I ran into while working on a side project, and I thought I’d share the idea + hack I came up with. Curious if anyone has tried something similar. The Problem On Vercel’s free plan, private repos auto-deploy only when there’s a new commit by the repo owner. You can’t manually trigger a deploy for a private repo. If a collaborator pushes commits, those changes won’t be deployed unless the repo owner also pushes something. The current workaround is trivial: I usually just add a fake commit like changing a character in the README.md, which triggers the pipeline and deploys the actual code. Annoying and manual. Solution (Source Code (https://github.com/satvikprsd/AutoBot)) I built a small Node.js server that: Listens to GitHub webhooks (push events). If someone else pushes code, the server appends a log line to auto_deploy_log.txt with a timestamp + author. The server then commits & pushes that trivial change using repo owner's account (using github token). Vercel sees a new commit → boom, auto-deploy triggered, no manual step needed. Would love any feedback on this. <!-- SC_ON --> submitted by /u/Deathfile78 (https://www.reddit.com/user/Deathfile78)
[link] (https://github.com/satvikprsd/AutoBot) [comments] (https://www.reddit.com/r/programming/comments/1npxaag/automating_vercel_deploys_on_private_repos/)
A step by step guide on how to build a LLM from scratch
https://www.reddit.com/r/programming/comments/1nq0166/a_step_by_step_guide_on_how_to_build_a_llm_from/

<!-- SC_OFF -->I wanted to share this here and hopefully it will help some folks to get deeper in this and help learn. I just published a comprehensive guide on how to build a LLM from scratch using historical London texts from 1500-1850. What I Built: Two identical models (117M & 354M parameters) trained from scratch Custom historical tokenizer with 30k vocabulary + 150+ special tokens for archaic English Complete data pipeline processing 218+ historical sources (500M+ characters) Production-ready training with multi-GPU support, WandB integration, and checkpointing Published models on Hugging Face ready for immediate use Why This Matters: Most LLM guides focus on fine-tuning existing models. This series shows you how to build from the ground up—eliminating modern biases and creating models that truly understand historical language patterns, cultural contexts, and period-specific knowledge. Resources: Blog Series: https://blog.desigeek.com/post/2025/09/building-llm-from-scratch-part1/ Complete Codebase: https://github.com/bahree/helloLondon Published Models: https://huggingface.co/bahree/london-historical-slm LinkedIn (if that's your thing): https://www.linkedin.com/feed/update/urn:li:share:7376863225306365952/ The models are already working and generating authentic 18th-century London text. Perfect for developers who want to understand the complete LLM development pipeline. Shoutout: Big thanks to u/Remarkable-Trick-177 (https://www.reddit.com/user/Remarkable-Trick-177/) for the inspiration! <!-- SC_ON --> submitted by /u/amitbahree (https://www.reddit.com/user/amitbahree)
[link] (https://blog.desigeek.com/post/2025/09/building-llm-from-scratch-part1/) [comments] (https://www.reddit.com/r/programming/comments/1nq0166/a_step_by_step_guide_on_how_to_build_a_llm_from/)
Table sorting
https://www.reddit.com/r/programming/comments/1nq05mi/table_sorting/

<!-- SC_OFF -->Yes, that simple table sorting. 10 years ago when I started my career that was the "take home" assignment. Today after trying to sort some simple values from a website I am amazed this problem was not solved yet. Just include the god damn sorting in the HTML spec and be done with it. Every table everywhere gets sort capabilities without coding. Thanks for reading my 3AM rant. <!-- SC_ON --> submitted by /u/FrostyCartoonist8523 (https://www.reddit.com/user/FrostyCartoonist8523)
[link] (https://localhost.com/) [comments] (https://www.reddit.com/r/programming/comments/1nq05mi/table_sorting/)
Create systems of equations and basic algebra app
https://www.reddit.com/r/programming/comments/1nq0tif/create_systems_of_equations_and_basic_algebra_app/

<!-- SC_OFF -->I want to create an app to 1. Use letters, Greek characters and subscripts for variable and equations. So typing “/Omega” will make Ω appear in its place. Perhaps there would be a panel of Greek characters I could click on as well 2. Input known variables 3. Input the equations apart of the system of equations 4. Automatically solve the system of equations 5. Add additional equations that will utilize the results of solving the system of equations. 6. Store equations I use over and over that I can quickly select. 7. The equations and values will need to be changeable at any point of the process I want the UI to be clean and the subscripts to actually look like subscripts when they are imputed and outputted. What language(s) should I use to create this? <!-- SC_ON --> submitted by /u/SkiMtVidGame-aineer (https://www.reddit.com/user/SkiMtVidGame-aineer)
[link] (http://thisisnotareallink.com/) [comments] (https://www.reddit.com/r/programming/comments/1nq0tif/create_systems_of_equations_and_basic_algebra_app/)