I need to store documents such as .pdf, .doc and .txt files to MaprDB. I saw one example in Hbase where it stores files in binary and is retrieved as files in Hue, but I not sure how it could be implemented. Any idea how can a document be stored in MaprDB?
Store documents (.pdf, .doc and .txt files) in MaprDB
331 views Asked by Amu At
1
There are 1 answers
Related Questions in JAVA
- I need the BIRT.war that is compatible with Java 17 and Tomcat 10
- Creating global Class holder
- No method found for class java.lang.String in Kafka
- Issue edit a jtable with a pictures
- getting error when trying to launch kotlin jar file that use supabase "java.lang.NoClassDefFoundError"
- Does the && (logical AND) operator have a higher precedence than || (logical OR) operator in Java?
- Mixed color rendering in a JTable
- HTTPS configuration in Spring Boot, server returning timeout
- How to use Layout to create textfields which dont increase in size?
- Function for making the code wait in javafx
- How to create beans of the same class for multiple template parameters in Spring
- How could you print a specific String from an array with the values of an array from a double array on the same line, using iteration to print all?
- org.telegram.telegrambots.meta.exceptions.TelegramApiException: Bot token and username can't be empty
- Accessing Secret Variables in Classic Pipelines through Java app in Azure DevOps
- Postgres && statement Error in Mybatis Mapper?
Related Questions in HBASE
- Apache atlas UI not showing up
- HBase Zookeeper Connection Error Docker Standalone 2.3.x and 2.4.x
- How does bulkload in databases such as hbase/cassandra/KV store work?
- How to eradicate the slowness caused due to reading rows from bigtable with hbase client in google dataflow job?
- i cant delete the specific column data by Timestamp
- hbase shell QualifierFilter is not filtering out columns when used with logical OR and SingleColumnValueFilter
- Spark - Fetch Hbase table all versions data using HBase Spark connector
- Unable to recover inconsistency in Hbase
- hBase java api, error on bulkload Added a key not lexically larger than previous sort (with JavaPairRDD<ImmutableBytesWritable, KeyValue>)
- Functionality inside completable future is completing quickly but completable future and timelimiter are taking too long
- about hbase put attribute
- java.lang.NoClassDefFoundError: org/apache/hadoop/hbase/client/Table
- Big Table Java Connectivity issue
- How to check if the Thrift is working on HBase version 2.5 and How to indicate if Thrift 1 or Thrift 2 is installed?
- HMaster stuck at "Initialize ServerManager and schedule SCP for crash servers"
Related Questions in HUE
- Hive query on HUE shows different timestamp than programatically/on data
- Podman Hue Container doesn't with Hadoop 3 ecosystem
- Counting the difference between two tables and selecting the different observations yield different results
- HUE AND KEYCLOAK INTEGRATION 403 Forbidden Error
- Replace specific color in HLSL shader
- Spark creating empty hive tables
- There is an issue on connecting to the DB in HUE
- HUE with Kerberized Hive
- How to hide null fields in hive(Hue, beeline)?
- searborn plot write legend on x-axis
- Connection is established hive wd Hue but unable to read data
- Hue does not work with https to connect Presto
- In hue notebook under sources hive is not showing up
- Unable to restart Apache hue running on docker container
- revoke invalidate metadata from hue users
Related Questions in MAPR
- ERROR: org.apache.hadoop.fs.UnsupportedFileSystemException: No FileSystem for scheme "maprfs"
- Is it possible to query MAPR hdfs/hive tables from Trino?
- Delete larger dataset with 10 Million records from HPE mapR table
- which MapR version is supported for Java17?
- java.lang.RuntimeException: Failure loading MapRClient
- How do we rename or copy the existing Hbase table in MapR without snapshot?
- Unable to connect to any of the cluster's CLDBs
- MapR Client for macos m1 chipset
- Configure Apache Drill to read xml files in Mapr distribution
- Map Reduce program to calculate the average and count
- Map function over nested columns
- hive tez mapr distribution query have issues with tez engine
- Removing special character in data in databricks
- How can I read files from a MapR cluster using Go?
- Hadoop configure cluter queried based on a flag/env parameter
Related Questions in NOSQL
- In Redis Databases how do we need to calculate the table size
- DynamoDB structure recommendation
- Efficiently read Firestore's document reference field contents
- Removing blocked users from the pipeline with lookup in mongodb
- Make a Cluster without using MongoDB Atlas
- MongoDB: Reading a large file vs uploading in a collection
- Mongo DB find objects (arrays) from Object
- Horizontal scaling strategy with 10,000 shards
- MongoDB aggregation - sum of array of nested objects
- how to configure mongodb to always cache 100% of a collection on RAM?
- Mongo Db global filter with C#
- TypeORM/MongoDB - sort collection
- Use Mongo $text search in limited set
- Not a value in projecting or not projecting MongoDB
- Which database management system should I use for this task?
Popular Questions
- How do I undo the most recent local commits in Git?
- How can I remove a specific item from an array in JavaScript?
- How do I delete a Git branch locally and remotely?
- Find all files containing a specific text (string) on Linux?
- How do I revert a Git repository to a previous commit?
- How do I create an HTML button that acts like a link?
- How do I check out a remote Git branch?
- How do I force "git pull" to overwrite local files?
- How do I list all files of a directory?
- How to check whether a string contains a substring in JavaScript?
- How do I redirect to another webpage?
- How can I iterate over rows in a Pandas DataFrame?
- How do I convert a String to an int in Java?
- Does Python have a string 'contains' substring method?
- How do I check if a string contains a specific word?
Trending Questions
- UIImageView Frame Doesn't Reflect Constraints
- Is it possible to use adb commands to click on a view by finding its ID?
- How to create a new web character symbol recognizable by html/javascript?
- Why isn't my CSS3 animation smooth in Google Chrome (but very smooth on other browsers)?
- Heap Gives Page Fault
- Connect ffmpeg to Visual Studio 2008
- Both Object- and ValueAnimator jumps when Duration is set above API LvL 24
- How to avoid default initialization of objects in std::vector?
- second argument of the command line arguments in a format other than char** argv or char* argv[]
- How to improve efficiency of algorithm which generates next lexicographic permutation?
- Navigating to the another actvity app getting crash in android
- How to read the particular message format in android and store in sqlite database?
- Resetting inventory status after order is cancelled
- Efficiently compute powers of X in SSE/AVX
- Insert into an external database using ajax and php : POST 500 (Internal Server Error)
First thing is , Im not aware about Maprdb as Im using Cloudera. But I have experience in hbase storing many types of objects in hbase as byte array like below mentioned.
Most primitive way of storing in hbase or any other db is byte array. see my answer
You can do that in below way using Apache commons lang API. probably this is best option, which will be applicable to all objects including image/audio/video etc..
please test this method with one of object type of any of your files.
SerializationUtils.serializewill return bytes. which you can insert.Note :jar of apache commons lang always available in hadoop cluster.(not external dependency)
another example :
For any reason if you don't want to use
SerializationUtilsclass provided by Apache commons lang, then you can see below pdf serialize and deserialize example for your better understanding but its lengthy code if you useSerializationUtilsthe code will be reduced.Above you are getting byte array you can prepare put request to upload to database i.e Hbase or any other database
Once you persisted, you can get the same using hbase get or
scanyougetyour pdf bytes and use the below code to again make same file i.e someFile.pdf in this case.EDIT : Since you asked HBASE examples I'm adding this.. in the below method
yourcolumnasBytearrayis your doc file for instance pdf.. converted to byte array (usingSerializationUtils.serialize) in above examples...