lama-cleaner全部命令行参数
2024-10-20 04:13:24
usage: main.py [-h] [--host HOST] [--port PORT] [--config-installer] [--load-installer-config] [--installer-config INSTALLER_CONFIG] [--model {lama,ldm,zits,mat,fcf,sd1.5,cv2,manga,sd2,paint_by_example,instruct_pix2pix}] [--no-half]
[--cpu-offload] [--disable-nsfw] [--sd-cpu-textencoder] [--local-files-only] [--enable-xformers] [--device {cuda,cpu,mps}] [--gui] [--no-gui-auto-close] [--gui-size GUI_SIZE GUI_SIZE] [--input INPUT]
[--output-dir OUTPUT_DIR] [--model-dir MODEL_DIR] [--disable-model-switch] optional arguments:
-h, --help show this help message and exit
--host HOST
--port PORT
--config-installer Open config web page, mainly for windows installer (default: False)
--load-installer-config
Load all cmd args from installer config file (default: False)
--installer-config INSTALLER_CONFIG
Config file for windows installer (default: None)
--model {lama,ldm,zits,mat,fcf,sd1.5,cv2,manga,sd2,paint_by_example,instruct_pix2pix}
--no-half Using full precision model. If your generate result is always black or green, use this argument. (sd/paint_by_exmaple) (default: False)
--cpu-offload Offloads all models to CPU, significantly reducing vRAM usage. (sd/paint_by_example) (default: False)
--disable-nsfw Disable NSFW checker. (sd/paint_by_example) (default: False)
--sd-cpu-textencoder Run Stable Diffusion text encoder model on CPU to save GPU memory. (default: False)
--local-files-only Use local files only, not connect to Hugging Face server. (sd/paint_by_example) (default: False)
--enable-xformers Enable xFormers optimizations. Requires xformers package has been installed. See: https://github.com/facebookresearch/xformers (sd/paint_by_example) (default: False)
--device {cuda,cpu,mps}
--gui Launch Lama Cleaner as desktop app (default: False)
--no-gui-auto-close Prevent backend auto close after the GUI window closed. (default: False)
--gui-size GUI_SIZE GUI_SIZE
Set window size for GUI (default: [1600, 1000])
--input INPUT If input is image, it will be loaded by default. If input is directory, you can browse and select image in file manager. (default: None)
--output-dir OUTPUT_DIR
Result images will be saved to output directory automatically without confirmation. (default: None)
--model-dir MODEL_DIR
Model download directory (by setting XDG_CACHE_HOME environment variable), by default model downloaded to ~/.cache (default: /Users/cwq/.cache)
--disable-model-switch
Disable model switch in frontend (default: False)
安装 Lama Cleaner 最简单的方法是通过(支持 python 3.7 ~ 3.10)安装它
# In order to use the GPU, install cuda version of pytorch first.
# pip install torch==1.13.1+cu117 --extra-index-url https://download.pytorch.org/whl/cu117
pip install lama-cleaner
最新文章
- [Spring MVC] - 500/404错误处理
- 基于HTML5的Web SCADA工控移动应用
- Noip2000 T3 单词接龙
- Meet Github
- SeismicPro地震剖面显示程序
- 项目管理和缺陷跟踪工具Redmine
- C#.NET ObjectDumper
- Apache Spark源码走读之9 -- Spark源码编译
- 【Qt】Qt之Tab键切换焦点顺序【转】
- ubuntu后台配置无线网络
- 安装 ArcGIS Runtime SDK for Android
- keybd_event函数用法
- vim阅读c++代码的快捷键
- 动态改变数据库连接 in Entity Framework 5
- HDOJ 2689
- java 空指针异常造成的原因有哪些
- lua State加载部分库
- GitChat招募IT类写作作者
- ActiveMQ系列之五:ActiveMQ的Transport
- maven中央仓库地址(支持db2,informix等)