<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Hacker News: dallen97</title><link>https://news.ycombinator.com/user?id=dallen97</link><description>Hacker News RSS</description><docs>https://hnrss.org/</docs><generator>hnrss v2.1.1</generator><lastBuildDate>Mon, 04 May 2026 00:07:47 +0000</lastBuildDate><atom:link href="https://hnrss.org/user?id=dallen97" rel="self" type="application/rss+xml"></atom:link><item><title><![CDATA[Show HN: Seedream 5.0: free AI image generator that claims strong text rendering]]></title><description><![CDATA[
<p>Found this today while looking for text to image tools that handle typography well.<p>Link: <a href="https://seedream5ai.org/" rel="nofollow">https://seedream5ai.org/</a><p>What it appears to offer
• Text to image generation from prompts, positioned for creators and designers   
• A focus on “text heavy graphics” and text rendering quality (based on how the site markets it)   
• Extra utilities in the nav like an image upscaler and background remover   
• A changelog page that describes a “My Images” workflow and an API endpoint for listing generations<p>Why it might be useful
• If you regularly generate posters, banners, thumbnails, UI mockups, or ads where readable text matters, it could be worth a quick test.<p>Questions for anyone who tries it
• How good is the actual text rendering versus other generators you’ve used
• Does it stay consistent with longer phrases and mixed language text
• Any issues with generation speed, pricing clarity, or output restrictions (watermark, resolution, etc.)<p>If you test it, please share prompt examples and results.</p>
<hr>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=46932496">https://news.ycombinator.com/item?id=46932496</a></p>
<p>Points: 1</p>
<p># Comments: 0</p>
]]></description><pubDate>Sun, 08 Feb 2026 08:34:52 +0000</pubDate><link>https://seedream5ai.org</link><dc:creator>dallen97</dc:creator><comments>https://news.ycombinator.com/item?id=46932496</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46932496</guid></item><item><title><![CDATA[Show HN: Seedance 2.0 AI video generator for creators and ecommerce]]></title><description><![CDATA[
<p>Came across this tool today and figured some folks here might want to kick the tires.<p>Link: <a href="https://seedance-2.net/" rel="nofollow">https://seedance-2.net/</a><p>What it appears to offer
• Text to video generation via a simple web UI
• The site claims up to 4K output and watermark free downloads on paid plans
• Pricing mentions batch generation, private generations, and API access on higher tiers<p>Why it might be interesting
• Low friction to test since it says you can start without signing up
• Could be useful for quick marketing clips, product showcases, or social content<p>Questions for anyone who tries it
• How’s prompt adherence and temporal consistency
• How does it compare with other web based video generators you’ve used
• Any red flags in quality, pricing, or terms<p>If you test it, would love to hear results and comparable alternatives.</p>
<hr>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=46930149">https://news.ycombinator.com/item?id=46930149</a></p>
<p>Points: 1</p>
<p># Comments: 0</p>
]]></description><pubDate>Sun, 08 Feb 2026 00:54:22 +0000</pubDate><link>https://seedance-2.net</link><dc:creator>dallen97</dc:creator><comments>https://news.ycombinator.com/item?id=46930149</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46930149</guid></item><item><title><![CDATA[Show HN: Kling VIDEO 3.0 released: 15-second AI video generation model]]></title><description><![CDATA[
<p>Kling just announced VIDEO 3.0 - a significant upgrade from their 2.6 and O1 models.<p>Key improvements:<p>*Extended duration:*
• Up to 15 seconds of continuous video (vs previous 5-10 seconds)
• Flexible duration ranging from 3-15 seconds
• Better for complex action sequences and scene development<p>*Unified multimodal approach:*
• Integrates text-to-video, image-to-video, reference-to-video
• Video modification and transformation in one model
• Native audio generation (synchronized with video)<p>*Two variants:*
• VIDEO 3.0 (upgraded from 2.6)
• VIDEO 3.0 Omni (upgraded from O1)<p>*Enhanced capabilities:*
• Improved subject consistency with reference-based generation
• Better prompt adherence and output stability
• More flexibility in storyboarding and shot control<p>This positions Kling competitively against:
- Runway Gen-4.5 ($95/month)
- Sora 2 (limited access)
- Veo 3.1 (Google)
- Grok Imagine (just topped rankings)<p>The 15-second duration is particularly interesting - enables more narrative storytelling vs the typical 5-second clips. Combined with native audio, this could change workflows for content creators.<p>Pricing isn't mentioned in the announcement. Previous Kling models ranged from $10-40/month, significantly cheaper than Runway.<p>Anyone have access to test this yet? Curious how the quality compares to Runway and Sora at this new duration.</p>
<hr>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=46835895">https://news.ycombinator.com/item?id=46835895</a></p>
<p>Points: 4</p>
<p># Comments: 2</p>
]]></description><pubDate>Sat, 31 Jan 2026 12:06:28 +0000</pubDate><link>https://kling3.net</link><dc:creator>dallen97</dc:creator><comments>https://news.ycombinator.com/item?id=46835895</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=46835895</guid></item><item><title><![CDATA[New comment by dallen97 in "Fei-Fei Li's Marble: 3D world model for spatially consistent virtual environment"]]></title><description><![CDATA[
<p>World Labs (Fei-Fei Li’s new startup) just opened Marble, a multimodal world model that generates persistent 3D worlds from text, images, video and 3D layouts.<p>The key idea: instead of generating a single object or clip, Marble outputs an entire spatially consistent environment that you can walk through, edit, grow and export as Gaussian splats / meshes / video for use in engines and VR. It also comes with AI-native editing tools and a hybrid 3D editor for blocking out geometry before refining details.<p>It’s being pitched as a step toward “spatial intelligence” — world models that reason about 3D space for creativity, robotics, and embodied agents. I’d love to hear from people who have tried it or who are building similar systems: what’s still missing for this to be production-ready?</p>
]]></description><pubDate>Thu, 13 Nov 2025 07:48:14 +0000</pubDate><link>https://news.ycombinator.com/item?id=45911966</link><dc:creator>dallen97</dc:creator><comments>https://news.ycombinator.com/item?id=45911966</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45911966</guid></item><item><title><![CDATA[Fei-Fei Li's Marble: 3D world model for spatially consistent virtual environment]]></title><description><![CDATA[
<p>Article URL: <a href="https://marble-ai.org">https://marble-ai.org</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=45911965">https://news.ycombinator.com/item?id=45911965</a></p>
<p>Points: 3</p>
<p># Comments: 2</p>
]]></description><pubDate>Thu, 13 Nov 2025 07:48:14 +0000</pubDate><link>https://marble-ai.org</link><dc:creator>dallen97</dc:creator><comments>https://news.ycombinator.com/item?id=45911965</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45911965</guid></item><item><title><![CDATA[New comment by dallen97 in "LoRA Model – LoRA-focused image generation hub that reduces "choice friction""]]></title><description><![CDATA[
<p>As a user who bounces between SD1.5/SDXL/FLUX LoRAs, my recurring pain points are: (1) compatibility (don’t mix architectures), (2) weight tuning (0.x vs 1.0 debates), and (3) preview/compare under fixed conditions. These show up constantly on Reddit.<p>LoRAModel positions itself as a LoRA-centric generation & training platform, with Flux LoRA compatibility noted on-site, a model gallery, and plans that include training credits. Having the LoRA context collected in one place helps me get to a “first decent result” faster (and keeps me from mixing base models by mistake).<p>What I liked as a user:
 • It nudges you to respect base-model compatibility before you waste time (SD1.5 vs SDXL vs FLUX).   
 • The flow aligns with the community’s with/without-LoRA testing habit; see common comparison workflows.   
 • Pricing/Refund/Privacy/TOS are public, which makes commercial use decisions easier.<p>Not affiliated; just sharing something that reduced friction for me. Link: <a href="https://loramodel.org/" rel="nofollow">https://loramodel.org/</a></p>
]]></description><pubDate>Sat, 08 Nov 2025 08:07:51 +0000</pubDate><link>https://news.ycombinator.com/item?id=45855010</link><dc:creator>dallen97</dc:creator><comments>https://news.ycombinator.com/item?id=45855010</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45855010</guid></item><item><title><![CDATA[LoRA Model – LoRA-focused image generation hub that reduces "choice friction"]]></title><description><![CDATA[
<p>Article URL: <a href="https://loramodel.org">https://loramodel.org</a></p>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=45855009">https://news.ycombinator.com/item?id=45855009</a></p>
<p>Points: 2</p>
<p># Comments: 1</p>
]]></description><pubDate>Sat, 08 Nov 2025 08:07:51 +0000</pubDate><link>https://loramodel.org</link><dc:creator>dallen97</dc:creator><comments>https://news.ycombinator.com/item?id=45855009</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45855009</guid></item><item><title><![CDATA[Show HN: HuMo AI – Multi-modal human-centric video generator (text+image+audio)]]></title><description><![CDATA[
<p>HuMo AI, a browser-based video studio that turns text, images, and audio into human-centric videos with strong subject consistency and lip-sync.<p>What it does
 • Text→Video with controllable motion and scene composition.
 • Image→Video to animate a still with natural movements and camera motion.
 • Audio-visual sync for speech-driven lip movement and rhythm-matched motion.
 • Multi-modal fusion: combine text + reference images + audio in one run.
 • Export-ready output: high-resolution (up to 4K), common aspect ratios.<p>Why
Creative teams often juggle multiple tools: one for T2V, another for in-between motion, a third for lip-sync. I wanted a single studio that keeps identity consistent and aligns visuals with audio—useful for product explainers, character spots, and quick social posts.<p>What’s different
 • Built around multi-modal conditioning rather than single-input T2V.
 • Emphasis on identity/subject preservation across the whole clip.
 • Frame-level audio alignment for more natural lips and motion.
 • Workflow extras like shot lists / playbooks to speed up iteration.</p>
<hr>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=45291152">https://news.ycombinator.com/item?id=45291152</a></p>
<p>Points: 1</p>
<p># Comments: 0</p>
]]></description><pubDate>Thu, 18 Sep 2025 15:49:22 +0000</pubDate><link>https://humo-ai.com</link><dc:creator>dallen97</dc:creator><comments>https://news.ycombinator.com/item?id=45291152</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45291152</guid></item><item><title><![CDATA[Show HN: Mii Maker – Free browser-based Mii avatar creator (no signup)]]></title><description><![CDATA[
<p>Hi HN — I built Mii Maker, a free, browser-based editor for creating Mii-style avatars. You can start instantly without sign-up and export images for profiles, socials, or docs.<p>What it does
 • No registration, free & unlimited usage; open the site and start creating.
 • Real-time preview while tweaking face/eyes/hair/clothes/accessories.
 • Export & share: save to your device (multiple formats) and use anywhere.
 • Works on desktop/tablet/phone with the same UI.<p>Why I built it
I often needed quick, playful avatars for profiles and team docs, but most tools had account gates or heavy installers. I wanted a lightweight web editor you can use in seconds.<p>What’s different
 • Starts instantly in the browser; keeps the flow fast for casual use.
 • Simple presets + fine controls so you can get “close enough” quickly.
 • Focused on export quality/usability rather than game-console import formats.<p>Limitations / notes
 • It’s an independent web tool and not affiliated with Nintendo. “Mii” refers to Nintendo’s customizable avatar concept introduced with the Wii; this project is a fan-style creator for general image use, not for importing into consoles.
 • Feature requests welcome (e.g., SVG export, batch variants, color palettes).</p>
<hr>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=45263680">https://news.ycombinator.com/item?id=45263680</a></p>
<p>Points: 1</p>
<p># Comments: 1</p>
]]></description><pubDate>Tue, 16 Sep 2025 15:34:59 +0000</pubDate><link>https://miimaker.org/</link><dc:creator>dallen97</dc:creator><comments>https://news.ycombinator.com/item?id=45263680</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45263680</guid></item><item><title><![CDATA[Show HN: Seedream4 – Multi-reference, in/out-painting, 4K export]]></title><description><![CDATA[
<p>Hi HN — I built Seedream4, a browser-based AI image generator & editor. You can generate from text or upload an image and do precise edits (background removal, in/out-painting, style transfer, upscaling) and export PNG/JPG/WebP. There are example prompts you can run right away, and uploads (PNG/JPG/WEBP up to 10MB) support higher-res results including 4K.<p>Why
I got tired of manual masking/retouch steps for marketing and product visuals. I wanted something that combines text-to-image with instruction-based edits so you can keep lighting/perspective intact and iterate fast.<p>What’s different
 • Instruction-based edits with local brushes/masks, without wrecking the rest of the scene.
 • Multi-reference consistency to keep brand colors/character identity coherent across shots.
 • Export-ready workflow (PNG/JPG/WebP, transparent backgrounds; 1:1, 16:9, 9:16, etc.).
 • No-signup demo: click an example or upload and try.<p>How it works (high level)
The web app sends your prompt/images to third-party model APIs (e.g. modern image-creation models) via my backend. I provide the UI, masking tools, queues, and credit metering. Per the privacy policy: prompts/images are processed by model providers and not retained longer than needed to deliver results.<p>Pricing
There’s a free tier to try core generation/editing. Paid plans add more credits, 4K export, priority processing; commercial usage rights are included in Pro.<p>Known limitations / looking for feedback
 • Text & tiny details still fail on some compositions.
 • Complex multi-object edits sometimes need an extra mask pass.
 • Throughput can vary under load.
I’d love feedback on UX, failure cases you hit, performance expectations, API/CLI needs, and whether the pricing/credit model feels fair.<p>Happy to answer anything about the design choices, queues, and the editor’s internals. Thanks for checking it out.</p>
<hr>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=45258728">https://news.ycombinator.com/item?id=45258728</a></p>
<p>Points: 1</p>
<p># Comments: 0</p>
]]></description><pubDate>Tue, 16 Sep 2025 06:29:50 +0000</pubDate><link>https://seedream4-ai.com</link><dc:creator>dallen97</dc:creator><comments>https://news.ycombinator.com/item?id=45258728</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45258728</guid></item><item><title><![CDATA[Show HN: Hunyuan 3D – Image → 3D with PBR]]></title><description><![CDATA[
<p>Built an interface for image → 3D with PBR materials. Exports GLB; typical runs take ~1–2 minutes depending on complexity. It’s powered by Tencent’s open-source Hunyuan3D 2.1 models (<a href="https://hunyuan-3d.org" rel="nofollow">https://hunyuan-3d.org</a>).</p>
<hr>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=45147570">https://news.ycombinator.com/item?id=45147570</a></p>
<p>Points: 2</p>
<p># Comments: 0</p>
]]></description><pubDate>Sat, 06 Sep 2025 08:24:20 +0000</pubDate><link>https://hunyuan-3d.org</link><dc:creator>dallen97</dc:creator><comments>https://news.ycombinator.com/item?id=45147570</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45147570</guid></item><item><title><![CDATA[New comment by dallen97 in "Gemini 2.5 Flash Image"]]></title><description><![CDATA[
<p>Been testing both Flux Krea and Nano Banana for image editing tasks. Honestly, they’re closer than people think. Flux Krea nails character consistency and semantic edits—especially when working with multi-turn prompts. Nano Banana (aka Gemini 2.5 Flash) feels snappier and handles multi-image composition surprisingly well.<p>I used this test site to benchmark them side by side. <a href="https://flux-krea.app/" rel="nofollow">https://flux-krea.app/</a>   Results? Comparable quality, but different strengths. Flux is better for precision edits and prompt adherence. Nano Banana shines in speed and creative blending. Pick your tool based on workflow, not hype.</p>
]]></description><pubDate>Fri, 29 Aug 2025 08:42:37 +0000</pubDate><link>https://news.ycombinator.com/item?id=45061688</link><dc:creator>dallen97</dc:creator><comments>https://news.ycombinator.com/item?id=45061688</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45061688</guid></item><item><title><![CDATA[Show HN: Wplace Pixel Art – Convert Images to Wplace-Compatible Pixel Art]]></title><description><![CDATA[
<p>We built a free browser tool for the wplace.live community that converts any image into pixel-perfect art using Wplace’s official 64-color palette.<p>Features<p>Exact color matching with Wplace’s palette (from lospec)<p>Highlights free vs paid pixels (Wplace has tiered pricing)<p>Real-time preview with grid overlays<p>Floyd–Steinberg dithering for smoother gradients<p>All processing is local – no uploads, no tracking<p>Why we made it Wplace is a collaborative pixel canvas, but placing art manually is slow and error-prone. This tool helps creators preview, plan, and optimize their art before placing it on the canvas.<p>Built with vanilla JS + canvas. No backend. Try it: <a href="https://wplacepixelart.com" rel="nofollow">https://wplacepixelart.com</a> Feedback welcome!</p>
<hr>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=45021631">https://news.ycombinator.com/item?id=45021631</a></p>
<p>Points: 3</p>
<p># Comments: 0</p>
]]></description><pubDate>Tue, 26 Aug 2025 02:28:39 +0000</pubDate><link>https://wplacepixelart.com/</link><dc:creator>dallen97</dc:creator><comments>https://news.ycombinator.com/item?id=45021631</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45021631</guid></item><item><title><![CDATA[Wplace Pixel Art – Convert Images to Wplace-Compatible Pixel Art]]></title><description><![CDATA[
<p>We built a free browser tool for the wplace.live community that converts any image into pixel-perfect art using Wplace’s official 64-color palette.<p>Features<p>Exact color matching with Wplace’s palette (from lospec)<p>Highlights free vs paid pixels (Wplace has tiered pricing)<p>Real-time preview with grid overlays<p>Floyd–Steinberg dithering for smoother gradients<p>All processing is local – no uploads, no tracking<p>Why we made it Wplace is a collaborative pixel canvas, but placing art manually is slow and error-prone. This tool helps creators preview, plan, and optimize their art before placing it on the canvas.<p>Built with vanilla JS + canvas. No backend. Try it: https://wplacepixelart.com Feedback welcome!</p>
<hr>
<p>Comments URL: <a href="https://news.ycombinator.com/item?id=45021564">https://news.ycombinator.com/item?id=45021564</a></p>
<p>Points: 3</p>
<p># Comments: 2</p>
]]></description><pubDate>Tue, 26 Aug 2025 02:19:24 +0000</pubDate><link>https://news.ycombinator.com/item?id=45021564</link><dc:creator>dallen97</dc:creator><comments>https://news.ycombinator.com/item?id=45021564</comments><guid isPermaLink="false">https://news.ycombinator.com/item?id=45021564</guid></item></channel></rss>