hdfs
Here are 542 public repositories matching this topic...
Write a Windows .BAT equivalent of travis_ci_helpers.sh
We need this .BAT file for running integration tests under Appveyor, which is unable to run our existing bash script.
Alternatively, rewrite that script in Python so we can use it under both Travis and Appveyor.
Originally posted by @mpenkov in RaRe-Technologies/smart_open#479 (comment)
Along with the tests:
/ibis/ibis/expr/api.py:318: DeprecationWarning: ibis.nanosecond is deprecated, use ibis.interval(nanoseconds=n) instead
warnings.warn(msg.format(name), DeprecationWarning)
/ibis/ibis/expr/api.py:318: DeprecationWarning: ibis.microsecond is deprecated, use ibis.interval(microseconds=n) instead
warnings.warn(msg.format(name), DeprecationWarning)
/ibis/i
Similar to how unix ls works, param could be -t
@stavrospapadopoulos is planning to do a full pass on the docs in the next development cycle to improve consistency, including:
- further clarification of the
capacitydefinition (ref #1167). - change the parameter from
capacity -> sparse_capacity - per @jakebolewski,
set_capacityneeds to return the dense capacity (e.g. for a HL api iterator over tiles)
We have multiple DC with different HDFS. Basically if you want to work like that you have to provide name node as parameter to snakebite.
Recently we faced the issue when folder with important data was deleted without moving to .Trash.
I investigated code a little bit.
Issue 1: So, 'skiptrash' configuration is not used by code, I assume documentation wasn't updated on this.
Issue 2: It looks like
Not clear in docs relation between tasks.max and number of consumers attached to topic partition
Hi we have 25 topics each topic having 2 partition , we have created connect config having topics.regex, so that connector consumes from all 25 topics with tasks.max set to 50 i.e(one unique consumer per partition) but when we describe the consumer group only two unique consumers are attached to 50 partition.
here's the config:
{
"name": "testConnectorfinalTest04",
"config": {
We use the connect ui in a Kubernetes setup where a sidecar of the connect-ui is notified when a new connect cluster joins. This sidecar updates then the caddy server configuration (mostly proxy settings).
Unfortunately, the caddy server does not restart automatically in case of config changes.
Proposal: use a tool like inotifywait that listens on changes of the caddy config and restart th
-
Updated
Jun 12, 2020 - Python
The documentation says you can use S3 as a file sink but gives zero details on how to do so. There is one line linking somewhere else but the link is broken.
These are the docs: http://divolte-releases.s3-website-eu-west-1.amazonaws.com/divolte-collector/0.9.0/userdoc/html/configuration.html
and this is the broken link: https://wiki.apache.org/hadoop/AmazonS3
Having stdout output and configuration through command line options will allow to use Storagetapper as a command line utility.
Document how CONFIG_TEMPLATE_PATH has to be used in the configs section of the svc.yml.
template: {{CONFIG_TEMPLATE_PATH}}/myconfig.yml
For the local yml test BasicServiceSpecTest.java [73] sets CONFIG_TEMPLATE_PATH, for the distribution it has to be set to frameworkname-scheduler in the env section of the marathon.json.mustache.
-
Updated
Apr 24, 2018 - Scala
-
Updated
Aug 6, 2019 - Java
Given the new key-value store event stream, it'd be nice to have something like:
$ skein kv events <application id> [options...]
where the process blocks, and logs the event stream to the console until interrupted. This would be useful for debugging, as well as demos.
-
Updated
Jan 8, 2020 - Jupyter Notebook
-
Updated
Aug 30, 2019 - Java
Assuming a valid hive-site.xml, it will be possible to determine the active hive warehouse HDFS directory and HiveServer2 and Metastore URIs.
From there we should be able to perform a directory analysis on the hive warehouse parent directory and then all HDFS locations that represent tables / partitions.
-
Updated
Jun 27, 2020 - Java
Potential users are confused about how Euphoria compares to Apache Beam and what its feature set is. Please create a page in the wiki describing the set of supported features (maybe along the lines of https://beam.apache.org/documentation/runners/capability-matrix/) and the set of feature not supported compared to Beam.
Contributes to #21.
The "Mirror of Linkedin's Camus" description on the repo homepage is wrong by most reasonable definitions of the word mirror. The content of the repos appear to be quite different.
Improve this page
Add a description, image, and links to the hdfs topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with the hdfs topic, visit your repo's landing page and select "manage topics."
Sponsors SeaweedFS via Patreon https://www.patreon.com/seaweedfs
Describe the bug
README#Installation-guide Step 3 contains command:
go get github.com/chrislusf/seaweedfs/weedIt builds weed here without compiler errors, but
weedbinary panicing same as in #1058 :