Skip to content
#

Computer vision

Computer vision is an interdisciplinary field that deals with how computers can be made to gain high-level understanding of digital images and videos.

Here are 6,647 public repositories matching this topic...

jina
Stubatiger
Stubatiger commented Nov 29, 2021

ENV

Python 3.9
jina 2.5.0

Describe the bug

If i try to dump an image blob to a io.bytesio object an error is thrown

from jina import Document
import io
d =  Document(uri='steam_data/image_store/8c/5b/8c5b265b9c533636.png')
output = io.BytesIO() 

(
    d
    .load_uri_to_image_blob()
    .dump_ima
label-studio
gluon-cv
JiaMingLin
JiaMingLin commented Aug 3, 2021

Hi,
I need to download the something-to-something and jester datasets. But the 20bn website "https://20bn.com" are not available for weeks, the error message is "503 Service Temporarily Unavailable".

I have already downloaded the video data of something-to-something v2, and I need the label dataset. For the Jester, I need both video and label data. Can someone share me the

willsmithorg
willsmithorg commented Dec 26, 2021

Could FeatureTools be implemented as an automated preprocessor to Autogluon, adding the ability to handle multi-entity problems (i.e. Data split across multiple normalised database tables)? So if you supply Autogluon with a list of Dataframes instead of a single Dataframe it would first invoke FeatureTools:

  • take the multiple Dataframes (entities) and try to auto-infer the relationship betwee
tanujjain
tanujjain commented Dec 4, 2020

WHAT

Make plot_duplicates output clearer.

WHY

The graph plotted by plot_duplicates output can be cluttered, especially when there are several duplicates found for an image. The titles of each image run into each other, which makes it hard to read the plot. Additionally, the titles can be hard to read, possibly due to a static dpi. An example can be seen below:

![imgdedup_plot_iss

ZhiyuanChen
ZhiyuanChen commented Jul 21, 2020

Well #77 didn't work for me while resuming from checkpoint_18.pth. The problem is when we resume, the model and optimizer passed in the restore_from function are suitable for epoch less than 10 (till backbone is not training) because the cfg.TRAIN.START_EPOCH is 0 (passed in build_opt_lr function just before restore_from) initially so this mismatches the optimizer after backbone start training. So