Skip to content
#

hadoop

Here are 2,111 public repositories matching this topic...

hanbaoan123
hanbaoan123 commented Feb 24, 2020

Issue Description

When I run the example cartpole with the default parameters, it can not converge to the max reward 200, I wonder what went wrong.
360截图20200224095510956

Version Information

Please indicate relevant versions, including, if relevant:

  • Deeplearning4j versi
presto
tdcmeehan
tdcmeehan commented Apr 21, 2020

We have a value_at_quantile function for QDigest types. Occasionally, it's useful to also retrieve a quantile given a value (say, for X, which percentile in the QDigest does it fall under).

The signature would look like the following:

quantile_at_value(qdigest(T), DOUBLE) -> T

Where T is one of DOUBLE, REAL or BIGINT.

Open Source Fast Scalable Machine Learning Platform For Smarter Applications: Deep Learning, Gradient Boosting & XGBoost, Random Forest, Generalized Linear Modeling (Logistic Regression, Elastic Net), K-Means, PCA, Stacked Ensembles, Automatic Machine Learning (AutoML), etc.

  • Updated Jun 4, 2020
  • Jupyter Notebook
yiheng
yiheng commented Jul 11, 2018

Spark 2.3 officially support run on kubernetes. While our guide of "Run on Kubernetes" is still based on a special version of Spark 2.2, which is out of date. We need to:

  1. update that document to Spark 2.3
  2. release the corresponding docker images.
kszucs
kszucs commented Jan 21, 2019

Along with the tests:

 /ibis/ibis/expr/api.py:318: DeprecationWarning: ibis.nanosecond is deprecated, use ibis.interval(nanoseconds=n) instead
    warnings.warn(msg.format(name), DeprecationWarning)
  /ibis/ibis/expr/api.py:318: DeprecationWarning: ibis.microsecond is deprecated, use ibis.interval(microseconds=n) instead
    warnings.warn(msg.format(name), DeprecationWarning)
  /ibis/i
garyelephant
garyelephant commented Mar 18, 2019

FAQ 1. Waterdrop开发者自己开发插件时,是否需要了解Waterdrop代码,是否需要把代码写到Waterdrop项目里?

开发者开发的插件,与waterdrop项目可以完全没有关系,不需要把你的插件代码写到waterdrop项目里面。
插件可以是一个完全独立的项目,在里面你用java,scala,maven,sbt,gradle,都随你。
这也是我们建议开发者开发插件的方式。


FAQ 2. 以集群模式(cluster)运行waterdrop,提示找不到:plugins.tar.gz

使用cluster模式提交前,需要您先执行如下命令:

# 备注: 预计下一个版本发布时,v1.2.3 我们会支持插件目录自动打包,无需再执行此命令。
tar zcvf plugins.tar.gz plugins

将插件目

450+ AWS, Hadoop, Cloud, Kafka, Docker, Elasticsearch, RabbitMQ, Redis, HBase, Solr, Cassandra, ZooKeeper, HDFS, Yarn, Hive, Presto, Drill, Impala, Consul, Spark, Jenkins, Travis CI, Git, MySQL, Linux, DNS, Whois, SSL Certs, Yum Security Updates, Kubernetes, Mesos, Riak, MongoDB, Memcached, Couchbase, CouchDB, Neo4j, Ambari, Cloudera, Hortonworks, MapR. Products: Attivio, Blue Talon, Datameer, H2O, WanDisco, Zaloni

  • Updated Jun 3, 2020
  • Python
dragon2611
dragon2611 commented Feb 1, 2020

Not sure if it's worth adding to the documentation but I found I had to add the following to the end of the fstab command otherwise the system would not boot cleanly.

,x-systemd.mount-timeout=30,_netdev

The line in fstab now looks like

mfsmount /mnt/mymnt fuse mfssubfolder=mydir,allow_other,x-systemd.mount-timeout=30,_netdev

I think it was trying to mount mooseFS before it had fini

50+ DockerHub public images for Docker & Kubernetes - Hadoop, Kafka, ZooKeeper, HBase, Cassandra, Solr, SolrCloud, Presto, Apache Drill, Nifi, Spark, Mesos, Consul, Riak, OpenTSDB, Jython, Advanced Nagios Plugins & DevOps Tools repos on Alpine, CentOS, Debian, Fedora, Ubuntu, Superset, H2O, Serf, Alluxio / Tachyon, FakeS3

  • Updated Jun 3, 2020
  • Shell
xccoder
xccoder commented Dec 6, 2019
  1. 安装linkis jobtypes
    按照官方安装文档进行自动化安装,执行sh install.sh最后一步报错:{"error":"Missing required parameter 'execid'."}。并没有看到文档中所说的“如果安装成功最后会打印:{"status":"success"}”,但是能在azkaban的/plugins/jobtypes目录下看到已经安装好的linkis任务插件。通过排查在安装脚本最后一步会去调用"curl http://azkaban_ip:executor_port/executor?action=reloadJobTypePlugins"进行插件的刷新。重启azkaban executor日志中看到已经加载了插件的信息 `INFO [JobTypeManager][Azkaban] Loaded jobtype linkis

Improve this page

Add a description, image, and links to the hadoop topic page so that developers can more easily learn about it.

Curate this topic

Add this topic to your repo

To associate your repository with the hadoop topic, visit your repo's landing page and select "manage topics."

Learn more

You can’t perform that action at this time.