Giter Club home page Giter Club logo

ai-comic-factory's People

Contributors

idoubi avatar jbilcke avatar jbilcke-hf avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

ai-comic-factory's Issues

How to run this locally on Windows?

Hi, I am more of an artist and less of a coder with only some basic Python experience and I was looking to deploy this on my local computer to experiment with AI comic art. What I did was I git cloned this project to my local folder, opened .env file and changed reference APIs and added API keys to all use OpenAI, but I don't know what to do from here. How do I actually run this program? Sorry if this is an irrelevant issue, I would really appreciate some help.

Returns incorrect content type for Inference Endpoints

I played around with my own Inference Endpoint using stabilityai/stable-diffusion-xl-base-1.0 and the Inference API. The second one worked, my own Endpoint did not. In fact, the wrong content type (test/plain) seems to be delivered at the Inference Endpoints if it is not requested in the header. Just add a "Accept": "image/png" to the request header here, rebuild and everything works fine again.

[Error] Unable to clone repo.

I get the following error:

C:\AI\git>git clone https://github.com/jbilcke-hf/ai-comic-factory
Cloning into 'ai-comic-factory'...
remote: Enumerating objects: 1027, done.
remote: Counting objects: 100% (198/198), done.
remote: Compressing objects: 100% (131/131), done.
Receiving objects:  80% (822/1027), 2.83 MiB | 5Receiving objects:  81% (832/1027), 2.83 MiB | 5Receiving objects:  82% (843/1027), 2.83 MiB | 5Receiving objects:  83% (853/1027), 2.83 MiB | 5Receiving objects:  84% (863/1027), 2.83 MiB | 5Receiving objects:  85% (873/1027), 2.83 MiB | 5Receiving objects:  86% (884/1027), 2.83 MiB | 5Receiving objects:  87% (894/1027), 2.83 MiB | 5Receiving objects:  88% (904/1027), 2.83 MiB | 5Receiving objects:  89% (915/1027), 2.83 MiB | 5Receiving objects:  90% (925/1027), 2.83 MiB | 5Receiving objects:  91% (935/1027), 2.83 MiB | 5Receiving objects:  92% (945/1027), 2.83 MiB | 5Receiving objects:  93% (956/1027), 2.83 MiB | 5remote: Total 1027 (delta 127), reused 121 (delta 64), pack-reused 829Receiving objects:  94% (9Receiving objects:  95% (976/1027), 2.83 MiB | 5Receiving objects:  96% (986/1027), 2.83 MiB | 5Receiving objects:  97% (997/1027), 2.83 MiB | 5Receiving objects:  98% (1007/1027), 2.83 MiB | Receiving objects:  99% (1017/1027), 2.83 MiB | Receiving objects: 100% (1027/1027), 2.83 MiB | Receiving objects: 100% (1027/1027), 3.13 MiB | 5.95 MiB/s, done.

Resolving deltas: 100% (543/543), done.
error: invalid path 'public/favicon/Icon?'
fatal: unable to checkout working tree
warning: Clone succeeded, but checkout failed.
You can inspect what was checked out with 'git status'
and retry with 'git restore --source=HEAD :/'

The resulting directory is empty.

Too many requests with Dalle3 as rendering engine

RENDERING_ENGINE="OPENAI"
NEXT_PUBLIC_ENABLE_RATE_LIMITER="true"
AUTH_OPENAI_API_KEY="sk-cxxxxxx"
RENDERING_OPENAI_API_BASE_URL="https://api.openai.com/v1"
RENDERING_OPENAI_API_MODEL="dall-e-3"

I am trying to start this project on my Macbook using openai both as LLM engine and as rendering engine.

This is my current .env setting, I start project with npm run dev and I can see UI has been rendered successfully. But when I type some prompts to frontend website and hit "GO" button, the backend command line shows me I have too many requests for OpenAI rendering engine.

I have tried to modify let delay = enableRateLimiter ? (1000 + (500 * panelIndex)) : 1000 to let delay = enableRateLimiter ? (70000 + (500 * panelIndex)) : 1000 in index.tsx file, however it still didn't work for me. @jbilcke I believe this is not expected, could you please take a look at this issue. Appriciate it if you could help~.

Creating more than just 4koma

Nice to see the problem of guided comic generation being tackled. There is a bit of a nit-pick: there are only 1 page and 4 panels, instead of a streamline from prompt to full comic chapter. StoryBird can generate whole stories, allow for narrative tuning and editing, and then generate images accordingly. what would be the major issue preventing having this type of feature? https://storybird.ai/

Side note: for comic generation there are already tools for dynamic panels, which is worth checking, since going from storyline to deciding on aspect ratio of panel outputs look strong https://shortbread.ai/

[Help] I am unable to run it successfully.

TypeError: fetch failed
at Object.fetch (node:internal/deps/undici/undici:11457:11)
at process.processTicksAndRejections (node:internal/process/task_queues:95:5) {
cause: ConnectTimeoutError: Connect Timeout Error
at onConnectTimeout (ai-comic-factory/node_modules/.pnpm/[email protected]_biqbaboplfbrettd7655fr4n2y/node_modules/next/dist/compiled/undici/index.js:1:82152)
at ai-comic-factory/node_modules/.pnpm/[email protected]_biqbaboplfbrettd7655fr4n2y/node_modules/next/dist/compiled/undici/index.js:1:81644
at Immediate._onImmediate (ai-comic-factory/node_modules/.pnpm/[email protected]_biqbaboplfbrettd7655fr4n2y/node_modules/next/dist/compiled/undici/index.js:1:82034)
at process.processImmediate (node:internal/timers:476:21)
at process.callbackTrampoline (node:internal/async_hooks:130:17) {
code: 'UND_ERR_CONNECT_TIMEOUT'
}
}

I am using the INFERENCE_API method, and I have configured everything in the .env file, but every time I enter the prompt and click "Go," it gives an error. Am I using the method incorrectly? To be straightforward, I've installed all the packages and performed the standard operations. Do I still need to start Docker?

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.