Giter Club home page Giter Club logo

photo2cartoon's Introduction

人像卡通化 (Photo to Cartoon)

中文版 | English Version

该项目为小视科技卡通肖像探索项目。您可使用微信扫描下方二维码或搜索“AI卡通秀”小程序体验卡通化效果。

也可以前往我们的ai开放平台进行在线体验:https://ai.minivision.cn/#/coreability/cartoon

技术交流QQ群:937627932

Updates

简介

人像卡通风格渲染的目标是,在保持原图像ID信息和纹理细节的同时,将真实照片转换为卡通风格的非真实感图像。我们的思路是,从大量照片/卡通数据中习得照片到卡通画的映射。一般而言,基于成对数据的pix2pix方法能达到较好的图像转换效果,但本任务的输入输出轮廓并非一一对应,例如卡通风格的眼睛更大、下巴更瘦;且成对的数据绘制难度大、成本较高,因此我们采用unpaired image translation方法来实现。

Unpaired image translation流派最经典方法是CycleGAN,但原始CycleGAN的生成结果往往存在较为明显的伪影且不稳定。近期的论文U-GAT-IT提出了一种归一化方法——AdaLIN,能够自动调节Instance Norm和Layer Norm的比重,再结合attention机制能够实现精美的人像日漫风格转换。

与夸张的日漫风不同,我们的卡通风格更偏写实,要求既有卡通画的简洁Q萌,又有明确的身份信息。为此我们增加了Face ID Loss,使用预训练的人脸识别模型提取照片和卡通画的ID特征,通过余弦距离来约束生成的卡通画。

此外,我们提出了一种Soft-AdaLIN(Soft Adaptive Layer-Instance Normalization)归一化方法,在反规范化时将编码器的均值方差(照片特征)与解码器的均值方差(卡通特征)相融合。

模型结构方面,在U-GAT-IT的基础上,我们在编码器之前和解码器之后各增加了2个hourglass模块,渐进地提升模型特征抽象和重建能力。

由于实验数据较为匮乏,为了降低训练难度,我们将数据处理成固定的模式。首先检测图像中的人脸及关键点,根据人脸关键点旋转校正图像,并按统一标准裁剪,再将裁剪后的头像输入人像分割模型去除背景。

Start

安装依赖库

项目所需的主要依赖库如下:

  • python 3.6
  • pytorch 1.4
  • tensorflow-gpu 1.14
  • face-alignment
  • dlib
  • onnxruntime

Clone:

git clone https://github.com/minivision-ai/photo2cartoon.git
cd ./photo2cartoon

下载资源

谷歌网盘 | 百度网盘 提取码:y2ch

  1. 人像卡通化预训练模型:photo2cartoon_weights.pt(20200504更新),存放在models路径下。
  2. 头像分割模型:seg_model_384.pb,存放在utils路径下。
  3. 人脸识别预训练模型:model_mobilefacenet.pth,存放在models路径下。(From: InsightFace_Pytorch
  4. 卡通画开源数据:cartoon_data,包含trainBtestB
  5. 人像卡通化onnx模型:photo2cartoon_weights.onnx 谷歌网盘,存放在models路径下。

测试

将一张测试照片(亚洲年轻女性)转换为卡通风格:

python test.py --photo_path ./images/photo_test.jpg --save_path ./images/cartoon_result.png

测试onnx模型

python test_onnx.py --photo_path ./images/photo_test.jpg --save_path ./images/cartoon_result.png

训练

1.数据准备

训练数据包括真实照片和卡通画像,为降低训练复杂度,我们对两类数据进行了如下预处理:

  • 检测人脸及关键点。
  • 根据关键点旋转校正人脸。
  • 将关键点边界框按固定的比例扩张并裁剪出人脸区域。
  • 使用人像分割模型将背景置白。

我们开源了204张处理后的卡通画数据,您还需准备约1000张人像照片(为匹配卡通数据,尽量使用亚洲年轻女性照片,人脸大小最好超过200x200像素),使用以下命令进行预处理:

python data_process.py --data_path YourPhotoFolderPath --save_path YourSaveFolderPath

将处理后的数据按照以下层级存放,trainAtestA中存放照片头像数据,trainBtestB中存放卡通头像数据。

├── dataset
    └── photo2cartoon
        ├── trainA
            ├── xxx.jpg
            ├── yyy.png
            └── ...
        ├── trainB
            ├── zzz.jpg
            ├── www.png
            └── ...
        ├── testA
            ├── aaa.jpg 
            ├── bbb.png
            └── ...
        └── testB
            ├── ccc.jpg 
            ├── ddd.png
            └── ...

2.训练

重新训练:

python train.py --dataset photo2cartoon

加载预训练参数:

python train.py --dataset photo2cartoon --pretrained_weights models/photo2cartoon_weights.pt

多GPU训练(仍建议使用batch_size=1,单卡训练):

python train.py --dataset photo2cartoon --batch_size 4 --gpu_ids 0 1 2 3

Q&A

Q:为什么开源的卡通化模型与小程序中的效果有差异?

A:开源模型的训练数据收集自互联网,为了得到更加精美的效果,我们在训练小程序中卡通化模型时,采用了定制的卡通画数据(200多张),且增大了输入分辨率。此外,小程序中的人脸特征提取器采用自研的识别模型,效果优于本项目使用的开源识别模型。

Q:如何选取效果最好的模型?

A:首先训练模型200k iterations,然后使用FID指标挑选出最优模型,最终挑选出的模型为迭代90k iterations时的模型。

Q:关于人脸特征提取模型。

A:实验中我们发现,使用自研的识别模型计算Face ID Loss训练效果远好于使用开源识别模型,若训练效果出现鲁棒性问题,可尝试将Face ID Loss权重置零。

Q:人像分割模型是否能用与分割半身像?

A:不能。该模型是针对本项目训练的专用模型,需先裁剪出人脸区域再输入。

Tips

我们开源的模型是基于亚洲年轻女性训练的,对于其他人群覆盖不足,您可根据使用场景自行收集相应人群的数据进行训练。我们的开放平台提供了能够覆盖各类人群的卡通化服务,您可前往体验。如有定制卡通风格需求请联系商务:18852075216。

参考

U-GAT-IT: Unsupervised Generative Attentional Networks with Adaptive Layer-Instance Normalization for Image-to-Image Translation [Paper][Code]

InsightFace_Pytorch

photo2cartoon's People

Contributors

chenxwh avatar hao-qiang avatar minivision-ailab avatar panxinmiao avatar tomguluson92 avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

photo2cartoon's Issues

判别器相关

作者,您好,您的工作做的非常棒,给了我很多启发。我想请问设置两个分别在layer=5和layer=7的判别器的用意是什么呢?另外这个参数的设置是否有什么经验可循?

运行test.py时报错

Traceback (most recent call last):
File "test.py", line 56, in
cartoon = c2p.inference(img)
File "test.py", line 43, in inference
cartoon = self.net(face)[0][0]
File "C:\ProgramData\Anaconda3\envs\gpu\lib\site-packages\torch\nn\modules\module.py", line 489, in call
result = self.forward(*input, **kwargs)
File "C:\Users\24238\Desktop\zhangting\photo2cartoon\models\networks.py", line 120, in forward
x = self.DecodeBlock1(x, content_features4, style_features)
File "C:\ProgramData\Anaconda3\envs\gpu\lib\site-packages\torch\nn\modules\module.py", line 489, in call
result = self.forward(*input, **kwargs)
File "C:\Users\24238\Desktop\zhangting\photo2cartoon\models\networks.py", line 296, in forward
out = self.norm1(out, content_features, style_features)
File "C:\ProgramData\Anaconda3\envs\gpu\lib\site-packages\torch\nn\modules\module.py", line 489, in call
result = self.forward(*input, **kwargs)
File "C:\Users\24238\Desktop\zhangting\photo2cartoon\models\networks.py", line 354, in forward
out = self.norm(x, soft_gamma, soft_beta)
File "C:\ProgramData\Anaconda3\envs\gpu\lib\site-packages\torch\nn\modules\module.py", line 489, in call
result = self.forward(*input, **kwargs)
File "C:\Users\24238\Desktop\zhangting\photo2cartoon\models\networks.py", line 366, in forward
in_mean, in_var = torch.mean(input, dim=[2, 3], keepdim=True), torch.var(input, dim=[2, 3], keepdim=True)
TypeError: var(): argument 'dim' must be int, not list

finetuning

how can I finetune the existing photo2cartoon weights, I loaded the weights but then the training starts from 0 iterations

关于后端部署

想问下你们小程序后端部署也是用的torch模型么 能简单说下部署小细节么 感谢

关于卡通画像的关键点检测问题

您好,感谢您分享这个有趣的项目,我关于卡通画像关键点检测那边有一些疑问
现在的关键点检测大多都是在现实图像里面的,貌似不同直接套用在卡通画像里面,想问一下这边你们是怎么处理的,有用到迁移学习什么的吗?还是自己标注数据呢?

TypeError on importing ResnetGenerator

/usr/bin/python test.py --photo_path ./images/photo_test.jpg --save_path ./images/cartoon_result.png
2020-05-05 19:57:44.377745: I tensorflow/stream_executor/platform/default/dso_loader.cc:44] Successfully opened dynamic library libcudart.so.10.2
Traceback (most recent call last):
File "test.py", line 5, in
from models import ResnetGenerator
File "/home/pankaj/Learning/photo2cartoon/models/init.py", line 2, in
from .UGATIT_sadalin_hourglass import UgatitSadalinHourglass
File "/home/pankaj/Learning/photo2cartoon/models/UGATIT_sadalin_hourglass.py", line 7, in
from utils import *
File "/home/pankaj/Learning/photo2cartoon/utils/init.py", line 1, in
from .preprocess import Preprocess
File "/home/pankaj/Learning/photo2cartoon/utils/preprocess.py", line 2, in
from .face_seg import FaceSeg
File "/home/pankaj/Learning/photo2cartoon/utils/face_seg.py", line 4, in
import tensorflow as tf
File "/usr/lib/python3.8/site-packages/tensorflow/init.py", line 41, in
from tensorflow.python.tools import module_util as _module_util
File "/usr/lib/python3.8/site-packages/tensorflow/python/init.py", line 64, in
from tensorflow.python.framework.framework_lib import * # pylint: disable=redefined-builtin
File "/usr/lib/python3.8/site-packages/tensorflow/python/framework/framework_lib.py", line 52, in
from tensorflow.python.framework.importer import import_graph_def
File "/usr/lib/python3.8/site-packages/tensorflow/python/framework/importer.py", line 28, in
from tensorflow.python.framework import function
File "/usr/lib/python3.8/site-packages/tensorflow/python/framework/function.py", line 37, in
from tensorflow.python.ops import resource_variable_ops
File "/usr/lib/python3.8/site-packages/tensorflow/python/ops/resource_variable_ops.py", line 41, in
from tensorflow.python.ops import gen_logging_ops
File "/usr/lib/python3.8/site-packages/tensorflow/python/ops/gen_logging_ops.py", line 312, in
def image_summary(tag, tensor, max_images=3, bad_color=_execute.make_tensor("""dtype: DT_UINT8 tensor_shape { dim { size: 4 } } int_val: 255 int_val: 0 int_val: 0 int_val: 255""", "bad_color"), name=None):
File "/usr/lib/python3.8/site-packages/tensorflow/python/eager/execute.py", line 229, in make_tensor
text_format.Merge(v, pb)
File "/usr/lib/python3.8/site-packages/google/protobuf/text_format.py", line 696, in Merge
return MergeLines(
File "/usr/lib/python3.8/site-packages/google/protobuf/text_format.py", line 770, in MergeLines
return parser.MergeLines(lines, message)
File "/usr/lib/python3.8/site-packages/google/protobuf/text_format.py", line 795, in MergeLines
self._ParseOrMerge(lines, message)
File "/usr/lib/python3.8/site-packages/google/protobuf/text_format.py", line 817, in _ParseOrMerge
self._MergeField(tokenizer, message)
File "/usr/lib/python3.8/site-packages/google/protobuf/text_format.py", line 942, in _MergeField
merger(tokenizer, message, field)
File "/usr/lib/python3.8/site-packages/google/protobuf/text_format.py", line 1105, in _MergeScalarField
setattr(message, field.name, value)
TypeError: descriptor to field 'tensorflow.TensorProto.dtype' doesn't apply to 'TensorProto' object

有办法提高训练速度么?

每10个iteration大概耗时16s,200k iteration耗时有点长,

一般gan的batch_size很多设置为1, 有其他方法提升训练速度吗?

面部定位网络下载出错

2020-04-30 15:42:14.563154: W tensorflow/stream_executor/platform/default/dso_loader.cc:55] Could not load dynamic library 'cudart64_101.dll'; dlerror: cudart64_101.dll not found
2020-04-30 15:42:14.571325: I tensorflow/stream_executor/cuda/cudart_stub.cc:29] Ignore above cudart dlerror if you do not have a GPU set up on your machine.
等待响应时间长,报错
urllib.error.ContentTooShortError
求解决方法!!!

test时的报错 IndexError

Traceback (most recent call last):
File "test.py", line 59, in
cartoon = c2p.inference(img)
File "test.py", line 38, in inference
mask = face_rgba[:, :, 3][:, :, np.newaxis].copy() / 255.
IndexError: index 3 is out of bounds for axis 2 with size 3

这个要怎么解决呢?谢谢你们

关于训练数据

您好,非常不错的工作。
由于开源模型基于女性数据训练,因此结果带有明显的女性特征,而小程序的结果泛化能力更强。在 Readme 中,提到内部训练数据是200多张定制化的卡通画,想请问一下,这200多张是包含了男、女总共的数据大小吗?

训练集

请问提供的模型训练集中只有女性的数据吗?输入男性的图片生成的图片都被女性化了

训练需要多大显存?

报错:Tried to allocate 20.00 MiB (GPU 0; 5.94 GiB total capacity; 5.22 GiB already allocated; 20.81 MiB free; 5.42 GiB reserved in total by PyTorch
是因为我的显存太小吗 T.T

for help:error with munmap_chunk

dear maintainer,
你好,当我测试的时候在face.seg的tf.session报错如下
*** Error in `/mnt/lustre/shanyuhao/.conda/envs/yh0524/bin/python': munmap_chunk(): invalid pointer: 0x00007f17d33bed40 ***
请问如何解决?
谢谢!
Screenshot from 2020-07-15 ### 13-13-51

使用test.py后,face alignment提示报错

我是Python新手,我在使用 test.py 时碰到了问题。一直卡在 self.fa = face_alignment.FaceAlignment(face_alignment.LandmarksType._2D, flip_input=False)这一句上,我猜测的原因为pytorch的模型没加载好。

粘贴的图形-2

粘贴的图形-1

mask分割出错

在另一台机器能跑出正确结果,但今天在另一台配置的机器上跑发现mask出不来。
发现`utils/face_seg.py`里output_op run出来的值并不是0-1之前,最大值有1.5300853e+26,非常奇怪,经过sigmoid的值怎么会大于1?

版本应该是没错的:
Python 3.6.9
tensorflow-gpu 1.14.0
感谢提供下排查思路!

关于训练的一些问题

您好我有两个问题:(1)如果重新训练一个模型,trainA和trainB数据集中大概有多少张图片才可以?
(2)一般要训练到多少代(iteations)才能基本达到给出的转换效果?

关于训练数据问题

你好,我想问下,trainA人脸照片和trainB动漫人脸, 两份数据是需要对应人吗(或者说 trainB里面都是trainA里人的动漫脸)?
比如 trainA 目录下的 a.jpg 和 trainB目录下的 a.jpg 是同一个人,因为我看你用了feature计算特征距离

转onnx报错

ONNX export failed on ATen operator var because torch.onnx.symbolic_opset9.var does not exist,有大佬遇见过吗

ModuleNotFoundError: No module named 'face_alignment'

I try to run it in google colab.
https://colab.research.google.com/drive/1kt3JpWFvmlmjv4lGQrKNNRKVoOqsw6uV?usp=sharing
But getting an error.

Traceback (most recent call last):
  File "test.py", line 5, in <module>
    from models import ResnetGenerator
  File "/content/photo2cartoon/models/__init__.py", line 2, in <module>
    from .UGATIT_sadalin_hourglass import UgatitSadalinHourglass
  File "/content/photo2cartoon/models/UGATIT_sadalin_hourglass.py", line 7, in <module>
    from utils import *
  File "/content/photo2cartoon/utils/__init__.py", line 1, in <module>
    from .preprocess import Preprocess
  File "/content/photo2cartoon/utils/preprocess.py", line 1, in <module>
    from .face_detect import FaceDetect
  File "/content/photo2cartoon/utils/face_detect.py", line 4, in <module>
    import face_alignment
ModuleNotFoundError: No module named 'face_alignment'

关于test

在window10环境中调用test.py:
python test.py --photo_path ./images/photo_test.jpg --save_path ./images/cartoon_result.png
程序没有报错,也无法得出结果。

Warning: this detector is deprecated. Please use a different one, i.e.: S3FD.
WARNING:tensorflow:From E:\DL_with_coding\photo2cartoon\utils\face_seg.py:13: The name tf.ConfigProto is deprecated. Please use tf.compat.v1.
ConfigProto instead.

WARNING:tensorflow:From E:\DL_with_coding\photo2cartoon\utils\face_seg.py:16: The name tf.Session is deprecated. Please use tf.compat.v1.Sess
ion instead.

2020-04-22 11:46:56.162121: I tensorflow/core/platform/cpu_feature_guard.cc:142] Your CPU supports instructions that this TensorFlow binary w
as not compiled to use: AVX AVX2
2020-04-22 11:46:56.167602: I tensorflow/stream_executor/platform/default/dso_loader.cc:42] Successfully opened dynamic library nvcuda.dll
2020-04-22 11:46:57.083446: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1640] Found device 0 with properties:
name: GeForce GTX 1060 with Max-Q Design major: 6 minor: 1 memoryClockRate(GHz): 1.3415
pciBusID: 0000:01:00.0
2020-04-22 11:46:57.090947: I tensorflow/stream_executor/platform/default/dlopen_checker_stub.cc:25] GPU libraries are statically linked, ski
p dlopen check.
2020-04-22 11:46:57.096467: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1763] Adding visible gpu devices: 0
2020-04-22 11:46:58.398108: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1181] Device interconnect StreamExecutor with strength 1 edge
matrix:
2020-04-22 11:46:58.403445: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1187] 0
2020-04-22 11:46:58.406841: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1200] 0: N
2020-04-22 11:46:58.410200: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1326] Created TensorFlow device (/job:localhost/replica:0/task
:0/device:GPU:0 with 4714 MB memory) -> physical GPU (device: 0, name: GeForce GTX 1060 with Max-Q Design, pci bus id: 0000:01:00.0, compute
capability: 6.1)
WARNING:tensorflow:From E:\DL_with_coding\photo2cartoon\utils\face_seg.py:26: FastGFile.init (from tensorflow.python.platform.gfile) is d
eprecated and will be removed in a future version.
Instructions for updating:
Use tf.gfile.GFile.
WARNING:tensorflow:From E:\DL_with_coding\photo2cartoon\utils\face_seg.py:27: The name tf.GraphDef is deprecated. Please use tf.compat.v1.Gra
phDef instead.

代码似乎有bug?

文件:https://github.com/minivision-ai/photo2cartoon/blob/master/models/networks.py
行:224-237
是不是应该是:

 skip1 = self.ConvBlock1_1(x)
down1 = F.avg_pool2d(x, 2)
down1 = self.ConvBlock1_2(down1)

skip2 = **self.ConvBlock2_1(down1)**
down2 = F.avg_pool2d(down1, 2)
down2 = **self.ConvBlock2_2(down2)**

skip3 = **self.ConvBlock3_1(down2)**
down3 = F.avg_pool2d(down2, 2)
down3 = **self.ConvBlock3_2(down3)**

skip4 = **self.ConvBlock4_1(down3)**
down4 = F.avg_pool2d(down3, 2)
down4 = **self.ConvBlock4_2(down4)**

请问能分享下loss曲线或数据吗?

您好,我现在训练的d_loss在1.0-2.0徘徊,g_loss在40-80徘徊,训练了8k左右个iteration。为了训练稳定,我把face_id loss的weight调成0了,其他权重保持默认值。想问下您的实验中最终的d_loss和g_loss长啥样吗?或者如果您有画loss曲线,能否分享下呢?感谢。

关于安装test

hi,你好,我在windows下进行安装测试,但是一直在:
Downloading the Face Alignment Network(FAN). Please wait...

然后下载中断,有什么其他的解决方式吗,比如我先下载这个FAN之类的

另外,你们小程序是输入512,然后训练的trainb部分也是只有200多张吗?

如何选取效果最好的模型 和 训练数据?

1.请问在README.md,您说"首先训练模型20万步,然后使用FID指标挑选出最优模型,最终挑选出的模型为迭代9万步时的模型。"
为什么要这样做而不是直接训练到代码里默认的100万步?以及如何用FID指标挑选呢?

2.请问小程序中演示的效果,您在训练时,真实照片和卡通画像各大约多少张呢?

3.之前看抖音的宝宝特效(比如把明星变成宝宝) 很火,就就很想尝试一下,请问用您的代码可以实现类似的效果吗?您对尝试这块,数据或代码上有什么建议吗?

恳请回复,万分感谢!

关于数据集

你好,在readme的Q&A中,你们描述小程序中的效果是使用了200张定制卡通图训练,这个数据量是不是太小了,还是说这200张定制的卡通图只是用来最后微调的?
此外,你们定制卡通图的时候应该是参照真实人脸图像找专业人士绘制的吧,这样的话是不是等于有了成对的训练数据,你们用成对的数据进行有监督学习么,这样对最后的效果会不会有提升呢?

关于模型训练的问题

您好,请问如果要训练自己想要风格的模型,训练数据的真人图片和卡通图片需要对应吗

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.