1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253545556575859606162636465666768697071727374757677787980818283848586878889909192939495969798991001011021031041051061071081091101111121131141151161171181191201211221231241251261271281291301311321331341351361371381391401411421431441451461471481491501511521531541551561571581591601611621631641651661671681691701711721731741751761771781791801811821831841851861871881891901911921931941951961971981992002012022032042052062072082092102112122132142152162172182192202212222232242252262272282292302312322332342352362372382392402412422432442452462472482492502512522532542552562572582592602612622632642652662672682692702712722732742752762772782792802812822832842852862872882892902912922932942952962972982993003013023033043053063073083093103113123133143153163173183193203213223233243253263273283293303313323333343353363373383393403413423433443453463473483493503513523533543553563573583593603613623633643653663673683693703713723733743753763773783793803813823833843853863873883893903913923933943953963973983994004014024034044054064074084094104114124134144154164174184194204214224234244254264274284294304314324334344354364374384394404414424434444454464474484494504514524534544554564574584594604614624634644654664674684694704714724734744754764774784794804814824834844854864874884894904914924934944954964974984995005015025035045055065075085095105115125135145155165175185195205215225235245255265275285295305315325335345355365375385395405415425435445455465475485495505515525535545555565575585595605615625635645655665675685695705715725735745755765775785795805815825835845855865875885895905915925935945955965975985996006016026036046056066076086096106116126136146156166176186196206216226236246256266276286296306316326336346356366376386396406416426436446456466476486496506516526536546556566576586596606616626636646656666676686696706716726736746756766776786796806816826836846856866876886896906916926936946956966976986997007017027037047057067077087097107117127137147157167177187197207217227237247257267277287297307317327337347357367377387397407417427437447457467477487497507517527537547557567577587597607617627637647657667677687697707717727737747757767777787797807817827837847857867877887897907917927937947957967977987998008018028038048058068078088098108118128138148158168178188198208218228238248258268278288298308318328338348358368378388398408418428438448458468478488498508518528538548558568578588598608618628638648658668678688698708718728738748758768778788798808818828838848858868878888898908918928938948958968978988999009019029039049059069079089099109119129139149159169179189199209219229239249259269279289299309319329339349359369379389399409419429439449459469479489499509519529539549559569579589599609619629639649659669679689699709719729739749759769779789799809819829839849859869879889899909919929939949959969979989991000100110021003100410051006100710081009101010111012101310141015101610171018101910201021102210231024102510261027102810291030103110321033103410351036103710381039104010411042104310441045104610471048104910501051105210531054105510561057105810591060106110621063106410651066106710681069107010711072107310741075107610771078107910801081108210831084108510861087108810891090109110921093109410951096109710981099110011011102110311041105110611071108110911101111111211131114111511161117111811191120112111221123112411251126112711281129113011311132113311341135113611371138113911401141114211431144114511461147114811491150115111521153115411551156115711581159116011611162116311641165116611671168116911701171117211731174117511761177117811791180118111821183118411851186118711881189119011911192119311941195119611971198119912001201120212031204120512061207120812091210121112121213121412151216121712181219122012211222122312241225122612271228122912301231123212331234123512361237123812391240124112421243124412451246124712481249125012511252125312541255125612571258125912601261126212631264126512661267126812691270127112721273127412751276127712781279128012811282128312841285128612871288128912901291129212931294129512961297129812991300130113021303130413051306130713081309131013111312131313141315131613171318131913201321132213231324132513261327132813291330133113321333133413351336133713381339134013411342134313441345134613471348134913501351135213531354135513561357135813591360136113621363136413651366136713681369137013711372137313741375137613771378137913801381138213831384138513861387138813891390139113921393139413951396139713981399140014011402140314041405140614071408140914101411141214131414141514161417141814191420142114221423142414251426142714281429143014311432143314341435143614371438143914401441144214431444144514461447144814491450145114521453145414551456145714581459146014611462146314641465146614671468146914701471147214731474147514761477147814791480148114821483148414851486148714881489149014911492149314941495149614971498149915001501150215031504150515061507150815091510151115121513151415151516151715181519152015211522152315241525152615271528152915301531153215331534153515361537153815391540154115421543154415451546154715481549155015511552155315541555155615571558155915601561156215631564156515661567156815691570157115721573157415751576157715781579158015811582158315841585158615871588158915901591159215931594159515961597159815991600160116021603160416051606160716081609161016111612161316141615161616171618161916201621162216231624162516261627162816291630163116321633163416351636163716381639164016411642164316441645164616471648164916501651165216531654165516561657165816591660166116621663166416651666166716681669167016711672167316741675167616771678167916801681168216831684168516861687168816891690169116921693169416951696169716981699170017011702170317041705170617071708170917101711171217131714171517161717171817191720172117221723172417251726172717281729173017311732173317341735173617371738173917401741174217431744174517461747174817491750175117521753175417551756175717581759176017611762176317641765176617671768176917701771177217731774177517761777177817791780178117821783178417851786178717881789179017911792179317941795179617971798179918001801180218031804180518061807180818091810181118121813181418151816181718181819182018211822182318241825182618271828182918301831183218331834183518361837183818391840184118421843184418451846184718481849185018511852185318541855185618571858185918601861186218631864186518661867186818691870187118721873187418751876187718781879188018811882188318841885188618871888188918901891189218931894189518961897189818991900190119021903190419051906190719081909191019111912191319141915191619171918191919201921192219231924192519261927192819291930193119321933193419351936193719381939194019411942194319441945194619471948194919501951195219531954195519561957195819591960196119621963196419651966196719681969197019711972197319741975197619771978197919801981198219831984198519861987198819891990199119921993199419951996199719981999200020012002200320042005200620072008200920102011201220132014201520162017201820192020202120222023202420252026202720282029203020312032203320342035203620372038203920402041204220432044204520462047204820492050205120522053205420552056205720582059206020612062206320642065206620672068206920702071207220732074207520762077207820792080208120822083208420852086208720882089209020912092209320942095209620972098209921002101210221032104210521062107210821092110211121122113 |
- <?xml version="1.0" encoding="iso-8859-1" standalone="no"?>
- <!-- Generated by the JDiff Javadoc doclet -->
- <!-- (http://www.jdiff.org) -->
- <!-- on Wed Aug 24 13:54:04 PDT 2016 -->
- <api
- xmlns:xsi='http://www.w3.org/2001/XMLSchema-instance'
- xsi:noNamespaceSchemaLocation='api.xsd'
- name="Apache Hadoop HDFS 2.7.2"
- jdversion="1.0.9">
- <!-- Command line arguments = -doclet org.apache.hadoop.classification.tools.IncludePublicAnnotationsJDiffDoclet -docletpath /Users/vinodkv/Workspace/eclipse-workspace/apache-git/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/hadoop-annotations.jar:/Users/vinodkv/Workspace/eclipse-workspace/apache-git/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/jdiff.jar -verbose -classpath /Users/vinodkv/Workspace/eclipse-workspace/apache-git/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/classes:/Users/vinodkv/Workspace/eclipse-workspace/apache-git/hadoop/hadoop-common-project/hadoop-annotations/target/hadoop-annotations-2.7.2.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_40.jdk/Contents/Home/lib/tools.jar:/Users/vinodkv/Workspace/eclipse-workspace/apache-git/hadoop/hadoop-common-project/hadoop-auth/target/hadoop-auth-2.7.2.jar:/Users/vinodkv/.m2/repository/org/slf4j/slf4j-api/1.7.10/slf4j-api-1.7.10.jar:/Users/vinodkv/.m2/repository/org/apache/httpcomponents/httpclient/4.2.5/httpclient-4.2.5.jar:/Users/vinodkv/.m2/repository/org/apache/httpcomponents/httpcore/4.2.5/httpcore-4.2.5.jar:/Users/vinodkv/.m2/repository/org/apache/directory/server/apacheds-kerberos-codec/2.0.0-M15/apacheds-kerberos-codec-2.0.0-M15.jar:/Users/vinodkv/.m2/repository/org/apache/directory/server/apacheds-i18n/2.0.0-M15/apacheds-i18n-2.0.0-M15.jar:/Users/vinodkv/.m2/repository/org/apache/directory/api/api-asn1-api/1.0.0-M20/api-asn1-api-1.0.0-M20.jar:/Users/vinodkv/.m2/repository/org/apache/directory/api/api-util/1.0.0-M20/api-util-1.0.0-M20.jar:/Users/vinodkv/.m2/repository/org/apache/zookeeper/zookeeper/3.4.6/zookeeper-3.4.6.jar:/Users/vinodkv/.m2/repository/jline/jline/0.9.94/jline-0.9.94.jar:/Users/vinodkv/.m2/repository/org/apache/curator/curator-framework/2.7.1/curator-framework-2.7.1.jar:/Users/vinodkv/Workspace/eclipse-workspace/apache-git/hadoop/hadoop-common-project/hadoop-common/target/hadoop-common-2.7.2.jar:/Users/vinodkv/.m2/repository/org/apache/commons/commons-math3/3.1.1/commons-math3-3.1.1.jar:/Users/vinodkv/.m2/repository/commons-httpclient/commons-httpclient/3.1/commons-httpclient-3.1.jar:/Users/vinodkv/.m2/repository/commons-net/commons-net/3.1/commons-net-3.1.jar:/Users/vinodkv/.m2/repository/commons-collections/commons-collections/3.2.2/commons-collections-3.2.2.jar:/Users/vinodkv/.m2/repository/javax/servlet/jsp/jsp-api/2.1/jsp-api-2.1.jar:/Users/vinodkv/.m2/repository/com/sun/jersey/jersey-json/1.9/jersey-json-1.9.jar:/Users/vinodkv/.m2/repository/org/codehaus/jettison/jettison/1.1/jettison-1.1.jar:/Users/vinodkv/.m2/repository/com/sun/xml/bind/jaxb-impl/2.2.3-1/jaxb-impl-2.2.3-1.jar:/Users/vinodkv/.m2/repository/javax/xml/bind/jaxb-api/2.2.2/jaxb-api-2.2.2.jar:/Users/vinodkv/.m2/repository/javax/xml/stream/stax-api/1.0-2/stax-api-1.0-2.jar:/Users/vinodkv/.m2/repository/javax/activation/activation/1.1/activation-1.1.jar:/Users/vinodkv/.m2/repository/org/codehaus/jackson/jackson-jaxrs/1.9.13/jackson-jaxrs-1.9.13.jar:/Users/vinodkv/.m2/repository/org/codehaus/jackson/jackson-xc/1.9.13/jackson-xc-1.9.13.jar:/Users/vinodkv/.m2/repository/net/java/dev/jets3t/jets3t/0.9.0/jets3t-0.9.0.jar:/Users/vinodkv/.m2/repository/com/jamesmurty/utils/java-xmlbuilder/0.4/java-xmlbuilder-0.4.jar:/Users/vinodkv/.m2/repository/commons-configuration/commons-configuration/1.6/commons-configuration-1.6.jar:/Users/vinodkv/.m2/repository/commons-digester/commons-digester/1.8/commons-digester-1.8.jar:/Users/vinodkv/.m2/repository/commons-beanutils/commons-beanutils/1.7.0/commons-beanutils-1.7.0.jar:/Users/vinodkv/.m2/repository/commons-beanutils/commons-beanutils-core/1.8.0/commons-beanutils-core-1.8.0.jar:/Users/vinodkv/.m2/repository/org/apache/avro/avro/1.7.4/avro-1.7.4.jar:/Users/vinodkv/.m2/repository/com/thoughtworks/paranamer/paranamer/2.3/paranamer-2.3.jar:/Users/vinodkv/.m2/repository/org/xerial/snappy/snappy-java/1.0.4.1/snappy-java-1.0.4.1.jar:/Users/vinodkv/.m2/repository/com/google/code/gson/gson/2.2.4/gson-2.2.4.jar:/Users/vinodkv/.m2/repository/com/jcraft/jsch/0.1.42/jsch-0.1.42.jar:/Users/vinodkv/.m2/repository/org/apache/curator/curator-client/2.7.1/curator-client-2.7.1.jar:/Users/vinodkv/.m2/repository/org/apache/curator/curator-recipes/2.7.1/curator-recipes-2.7.1.jar:/Users/vinodkv/.m2/repository/com/google/code/findbugs/jsr305/3.0.0/jsr305-3.0.0.jar:/Users/vinodkv/.m2/repository/org/apache/commons/commons-compress/1.4.1/commons-compress-1.4.1.jar:/Users/vinodkv/.m2/repository/org/tukaani/xz/1.0/xz-1.0.jar:/Users/vinodkv/.m2/repository/com/google/guava/guava/11.0.2/guava-11.0.2.jar:/Users/vinodkv/.m2/repository/org/mortbay/jetty/jetty/6.1.26/jetty-6.1.26.jar:/Users/vinodkv/.m2/repository/org/mortbay/jetty/jetty-util/6.1.26/jetty-util-6.1.26.jar:/Users/vinodkv/.m2/repository/com/sun/jersey/jersey-core/1.9/jersey-core-1.9.jar:/Users/vinodkv/.m2/repository/com/sun/jersey/jersey-server/1.9/jersey-server-1.9.jar:/Users/vinodkv/.m2/repository/asm/asm/3.2/asm-3.2.jar:/Users/vinodkv/.m2/repository/commons-cli/commons-cli/1.2/commons-cli-1.2.jar:/Users/vinodkv/.m2/repository/commons-codec/commons-codec/1.4/commons-codec-1.4.jar:/Users/vinodkv/.m2/repository/commons-io/commons-io/2.4/commons-io-2.4.jar:/Users/vinodkv/.m2/repository/commons-lang/commons-lang/2.6/commons-lang-2.6.jar:/Users/vinodkv/.m2/repository/commons-logging/commons-logging/1.1.3/commons-logging-1.1.3.jar:/Users/vinodkv/.m2/repository/commons-daemon/commons-daemon/1.0.13/commons-daemon-1.0.13.jar:/Users/vinodkv/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar:/Users/vinodkv/.m2/repository/com/google/protobuf/protobuf-java/2.5.0/protobuf-java-2.5.0.jar:/Users/vinodkv/.m2/repository/javax/servlet/servlet-api/2.5/servlet-api-2.5.jar:/Users/vinodkv/.m2/repository/org/slf4j/slf4j-log4j12/1.7.10/slf4j-log4j12-1.7.10.jar:/Users/vinodkv/.m2/repository/org/codehaus/jackson/jackson-core-asl/1.9.13/jackson-core-asl-1.9.13.jar:/Users/vinodkv/.m2/repository/org/codehaus/jackson/jackson-mapper-asl/1.9.13/jackson-mapper-asl-1.9.13.jar:/Users/vinodkv/.m2/repository/xmlenc/xmlenc/0.52/xmlenc-0.52.jar:/Users/vinodkv/.m2/repository/io/netty/netty/3.6.2.Final/netty-3.6.2.Final.jar:/Users/vinodkv/.m2/repository/io/netty/netty-all/4.0.23.Final/netty-all-4.0.23.Final.jar:/Users/vinodkv/.m2/repository/xerces/xercesImpl/2.9.1/xercesImpl-2.9.1.jar:/Users/vinodkv/.m2/repository/xml-apis/xml-apis/1.3.04/xml-apis-1.3.04.jar:/Users/vinodkv/.m2/repository/org/apache/htrace/htrace-core/3.1.0-incubating/htrace-core-3.1.0-incubating.jar:/Users/vinodkv/.m2/repository/org/fusesource/leveldbjni/leveldbjni-all/1.8/leveldbjni-all-1.8.jar -sourcepath /Users/vinodkv/Workspace/eclipse-workspace/apache-git/hadoop/hadoop-hdfs-project/hadoop-hdfs/src/main/java -doclet org.apache.hadoop.classification.tools.IncludePublicAnnotationsJDiffDoclet -docletpath /Users/vinodkv/Workspace/eclipse-workspace/apache-git/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/hadoop-annotations.jar:/Users/vinodkv/Workspace/eclipse-workspace/apache-git/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/jdiff.jar -apidir /Users/vinodkv/Workspace/eclipse-workspace/apache-git/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/site/jdiff/xml -apiname Apache Hadoop HDFS 2.7.2 -->
- <package name="org.apache.hadoop.fs">
- <!-- start class org.apache.hadoop.fs.BlockStorageLocation -->
- <class name="BlockStorageLocation" extends="org.apache.hadoop.fs.BlockLocation"
- abstract="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <constructor name="BlockStorageLocation" type="org.apache.hadoop.fs.BlockLocation, org.apache.hadoop.fs.VolumeId[]"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <exception name="IOException" type="java.io.IOException"/>
- </constructor>
- <method name="getVolumeIds" return="org.apache.hadoop.fs.VolumeId[]"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[Gets the list of {@link VolumeId} corresponding to the block's replicas.
- @return volumeIds list of VolumeId for the block's replicas]]>
- </doc>
- </method>
- <doc>
- <![CDATA[Wrapper for {@link BlockLocation} that also adds {@link VolumeId} volume
- location information for each replica.]]>
- </doc>
- </class>
- <!-- end class org.apache.hadoop.fs.BlockStorageLocation -->
- <!-- start class org.apache.hadoop.fs.CacheFlag -->
- <class name="CacheFlag" extends="java.lang.Enum"
- abstract="false"
- static="false" final="true" visibility="public"
- deprecated="not deprecated">
- <method name="values" return="org.apache.hadoop.fs.CacheFlag[]"
- abstract="false" native="false" synchronized="false"
- static="true" final="false" visibility="public"
- deprecated="not deprecated">
- </method>
- <method name="valueOf" return="org.apache.hadoop.fs.CacheFlag"
- abstract="false" native="false" synchronized="false"
- static="true" final="false" visibility="public"
- deprecated="not deprecated">
- <param name="name" type="java.lang.String"/>
- </method>
- <doc>
- <![CDATA[Specifies semantics for CacheDirective operations. Multiple flags can
- be combined in an EnumSet.]]>
- </doc>
- </class>
- <!-- end class org.apache.hadoop.fs.CacheFlag -->
- <!-- start class org.apache.hadoop.fs.HdfsVolumeId -->
- <class name="HdfsVolumeId" extends="java.lang.Object"
- abstract="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <implements name="org.apache.hadoop.fs.VolumeId"/>
- <constructor name="HdfsVolumeId" type="byte[]"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- </constructor>
- <method name="compareTo" return="int"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <param name="arg0" type="org.apache.hadoop.fs.VolumeId"/>
- </method>
- <method name="hashCode" return="int"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- </method>
- <method name="equals" return="boolean"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <param name="obj" type="java.lang.Object"/>
- </method>
- <method name="toString" return="java.lang.String"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- </method>
- <doc>
- <![CDATA[HDFS-specific volume identifier which implements {@link VolumeId}. Can be
- used to differentiate between the data directories on a single datanode. This
- identifier is only unique on a per-datanode basis.]]>
- </doc>
- </class>
- <!-- end class org.apache.hadoop.fs.HdfsVolumeId -->
- <!-- start interface org.apache.hadoop.fs.VolumeId -->
- <interface name="VolumeId" abstract="true"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <implements name="java.lang.Comparable"/>
- <method name="compareTo" return="int"
- abstract="true" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <param name="arg0" type="org.apache.hadoop.fs.VolumeId"/>
- </method>
- <method name="hashCode" return="int"
- abstract="true" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- </method>
- <method name="equals" return="boolean"
- abstract="true" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <param name="obj" type="java.lang.Object"/>
- </method>
- <doc>
- <![CDATA[Opaque interface that identifies a disk location. Subclasses
- should implement {@link Comparable} and override both equals and hashCode.]]>
- </doc>
- </interface>
- <!-- end interface org.apache.hadoop.fs.VolumeId -->
- </package>
- <package name="org.apache.hadoop.hdfs">
- <!-- start class org.apache.hadoop.hdfs.DFSInotifyEventInputStream -->
- <class name="DFSInotifyEventInputStream" extends="java.lang.Object"
- abstract="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <method name="poll" return="org.apache.hadoop.hdfs.inotify.EventBatch"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <exception name="IOException" type="java.io.IOException"/>
- <exception name="MissingEventsException" type="org.apache.hadoop.hdfs.inotify.MissingEventsException"/>
- <doc>
- <![CDATA[Returns the next batch of events in the stream or null if no new
- batches are currently available.
- @throws IOException because of network error or edit log
- corruption. Also possible if JournalNodes are unresponsive in the
- QJM setting (even one unresponsive JournalNode is enough in rare cases),
- so catching this exception and retrying at least a few times is
- recommended.
- @throws MissingEventsException if we cannot return the next batch in the
- stream because the data for the events (and possibly some subsequent
- events) has been deleted (generally because this stream is a very large
- number of transactions behind the current state of the NameNode). It is
- safe to continue reading from the stream after this exception is thrown
- The next available batch of events will be returned.]]>
- </doc>
- </method>
- <method name="getTxidsBehindEstimate" return="long"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[Return a estimate of how many transaction IDs behind the NameNode's
- current state this stream is. Clients should periodically call this method
- and check if its result is steadily increasing, which indicates that they
- are falling behind (i.e. transaction are being generated faster than the
- client is reading them). If a client falls too far behind events may be
- deleted before the client can read them.
- <p/>
- A return value of -1 indicates that an estimate could not be produced, and
- should be ignored. The value returned by this method is really only useful
- when compared to previous or subsequent returned values.]]>
- </doc>
- </method>
- <method name="poll" return="org.apache.hadoop.hdfs.inotify.EventBatch"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <param name="time" type="long"/>
- <param name="tu" type="java.util.concurrent.TimeUnit"/>
- <exception name="IOException" type="java.io.IOException"/>
- <exception name="InterruptedException" type="java.lang.InterruptedException"/>
- <exception name="MissingEventsException" type="org.apache.hadoop.hdfs.inotify.MissingEventsException"/>
- <doc>
- <![CDATA[Returns the next event batch in the stream, waiting up to the specified
- amount of time for a new batch. Returns null if one is not available at the
- end of the specified amount of time. The time before the method returns may
- exceed the specified amount of time by up to the time required for an RPC
- to the NameNode.
- @param time number of units of the given TimeUnit to wait
- @param tu the desired TimeUnit
- @throws IOException see {@link DFSInotifyEventInputStream#poll()}
- @throws MissingEventsException
- see {@link DFSInotifyEventInputStream#poll()}
- @throws InterruptedException if the calling thread is interrupted]]>
- </doc>
- </method>
- <method name="take" return="org.apache.hadoop.hdfs.inotify.EventBatch"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <exception name="IOException" type="java.io.IOException"/>
- <exception name="InterruptedException" type="java.lang.InterruptedException"/>
- <exception name="MissingEventsException" type="org.apache.hadoop.hdfs.inotify.MissingEventsException"/>
- <doc>
- <![CDATA[Returns the next batch of events in the stream, waiting indefinitely if
- a new batch is not immediately available.
- @throws IOException see {@link DFSInotifyEventInputStream#poll()}
- @throws MissingEventsException see
- {@link DFSInotifyEventInputStream#poll()}
- @throws InterruptedException if the calling thread is interrupted]]>
- </doc>
- </method>
- <field name="LOG" type="org.slf4j.Logger"
- transient="false" volatile="false"
- static="true" final="false" visibility="public"
- deprecated="not deprecated">
- </field>
- <doc>
- <![CDATA[Stream for reading inotify events. DFSInotifyEventInputStreams should not
- be shared among multiple threads.]]>
- </doc>
- </class>
- <!-- end class org.apache.hadoop.hdfs.DFSInotifyEventInputStream -->
- <!-- start class org.apache.hadoop.hdfs.UnknownCipherSuiteException -->
- <class name="UnknownCipherSuiteException" extends="java.io.IOException"
- abstract="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <constructor name="UnknownCipherSuiteException" type="java.lang.String"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- </constructor>
- <doc>
- <![CDATA[Thrown when an unknown cipher suite is encountered.]]>
- </doc>
- </class>
- <!-- end class org.apache.hadoop.hdfs.UnknownCipherSuiteException -->
- <!-- start class org.apache.hadoop.hdfs.UnknownCryptoProtocolVersionException -->
- <class name="UnknownCryptoProtocolVersionException" extends="java.io.IOException"
- abstract="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <constructor name="UnknownCryptoProtocolVersionException"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- </constructor>
- <constructor name="UnknownCryptoProtocolVersionException" type="java.lang.String"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- </constructor>
- </class>
- <!-- end class org.apache.hadoop.hdfs.UnknownCryptoProtocolVersionException -->
- <doc>
- <![CDATA[<p>A distributed implementation of {@link
- org.apache.hadoop.fs.FileSystem}. This is loosely modelled after
- Google's <a href="http://research.google.com/archive/gfs.html">GFS</a>.</p>
- <p>The most important difference is that unlike GFS, Hadoop DFS files
- have strictly one writer at any one time. Bytes are always appended
- to the end of the writer's stream. There is no notion of "record appends"
- or "mutations" that are then checked or reordered. Writers simply emit
- a byte stream. That byte stream is guaranteed to be stored in the
- order written.</p>]]>
- </doc>
- </package>
- <package name="org.apache.hadoop.hdfs.client">
- <!-- start class org.apache.hadoop.hdfs.client.BlockReportOptions -->
- <class name="BlockReportOptions" extends="java.lang.Object"
- abstract="false"
- static="false" final="true" visibility="public"
- deprecated="not deprecated">
- <method name="isIncremental" return="boolean"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- </method>
- <method name="toString" return="java.lang.String"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- </method>
- <doc>
- <![CDATA[Options that can be specified when manually triggering a block report.]]>
- </doc>
- </class>
- <!-- end class org.apache.hadoop.hdfs.client.BlockReportOptions -->
- <!-- start class org.apache.hadoop.hdfs.client.HdfsAdmin -->
- <class name="HdfsAdmin" extends="java.lang.Object"
- abstract="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <constructor name="HdfsAdmin" type="java.net.URI, org.apache.hadoop.conf.Configuration"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <exception name="IOException" type="java.io.IOException"/>
- <doc>
- <![CDATA[Create a new HdfsAdmin client.
- @param uri the unique URI of the HDFS file system to administer
- @param conf configuration
- @throws IOException in the event the file system could not be created]]>
- </doc>
- </constructor>
- <method name="setQuota"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <param name="src" type="org.apache.hadoop.fs.Path"/>
- <param name="quota" type="long"/>
- <exception name="IOException" type="java.io.IOException"/>
- <doc>
- <![CDATA[Set the namespace quota (count of files, directories, and sym links) for a
- directory.
- @param src the path to set the quota for
- @param quota the value to set for the quota
- @throws IOException in the event of error]]>
- </doc>
- </method>
- <method name="clearQuota"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <param name="src" type="org.apache.hadoop.fs.Path"/>
- <exception name="IOException" type="java.io.IOException"/>
- <doc>
- <![CDATA[Clear the namespace quota (count of files, directories and sym links) for a
- directory.
- @param src the path to clear the quota of
- @throws IOException in the event of error]]>
- </doc>
- </method>
- <method name="setSpaceQuota"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <param name="src" type="org.apache.hadoop.fs.Path"/>
- <param name="spaceQuota" type="long"/>
- <exception name="IOException" type="java.io.IOException"/>
- <doc>
- <![CDATA[Set the storage space quota (size of files) for a directory. Note that
- directories and sym links do not occupy storage space.
- @param src the path to set the space quota of
- @param spaceQuota the value to set for the space quota
- @throws IOException in the event of error]]>
- </doc>
- </method>
- <method name="clearSpaceQuota"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <param name="src" type="org.apache.hadoop.fs.Path"/>
- <exception name="IOException" type="java.io.IOException"/>
- <doc>
- <![CDATA[Clear the storage space quota (size of files) for a directory. Note that
- directories and sym links do not occupy storage space.
- @param src the path to clear the space quota of
- @throws IOException in the event of error]]>
- </doc>
- </method>
- <method name="setQuotaByStorageType"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <param name="src" type="org.apache.hadoop.fs.Path"/>
- <param name="type" type="org.apache.hadoop.fs.StorageType"/>
- <param name="quota" type="long"/>
- <exception name="IOException" type="java.io.IOException"/>
- <doc>
- <![CDATA[Set the quota by storage type for a directory. Note that
- directories and sym links do not occupy storage type quota.
- @param src the target directory to set the quota by storage type
- @param type the storage type to set for quota by storage type
- @param quota the value to set for quota by storage type
- @throws IOException in the event of error]]>
- </doc>
- </method>
- <method name="clearQuotaByStorageType"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <param name="src" type="org.apache.hadoop.fs.Path"/>
- <param name="type" type="org.apache.hadoop.fs.StorageType"/>
- <exception name="IOException" type="java.io.IOException"/>
- <doc>
- <![CDATA[Clear the space quota by storage type for a directory. Note that
- directories and sym links do not occupy storage type quota.
- @param src the target directory to clear the quota by storage type
- @param type the storage type to clear for quota by storage type
- @throws IOException in the event of error]]>
- </doc>
- </method>
- <method name="allowSnapshot"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <param name="path" type="org.apache.hadoop.fs.Path"/>
- <exception name="IOException" type="java.io.IOException"/>
- <doc>
- <![CDATA[Allow snapshot on a directory.
- @param path The path of the directory where snapshots will be taken.]]>
- </doc>
- </method>
- <method name="disallowSnapshot"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <param name="path" type="org.apache.hadoop.fs.Path"/>
- <exception name="IOException" type="java.io.IOException"/>
- <doc>
- <![CDATA[Disallow snapshot on a directory.
- @param path The path of the snapshottable directory.]]>
- </doc>
- </method>
- <method name="addCacheDirective" return="long"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <param name="info" type="org.apache.hadoop.hdfs.protocol.CacheDirectiveInfo"/>
- <param name="flags" type="java.util.EnumSet"/>
- <exception name="IOException" type="java.io.IOException"/>
- <doc>
- <![CDATA[Add a new CacheDirectiveInfo.
- @param info Information about a directive to add.
- @param flags {@link CacheFlag}s to use for this operation.
- @return the ID of the directive that was created.
- @throws IOException if the directive could not be added]]>
- </doc>
- </method>
- <method name="modifyCacheDirective"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <param name="info" type="org.apache.hadoop.hdfs.protocol.CacheDirectiveInfo"/>
- <param name="flags" type="java.util.EnumSet"/>
- <exception name="IOException" type="java.io.IOException"/>
- <doc>
- <![CDATA[Modify a CacheDirective.
- @param info Information about the directive to modify. You must set the ID
- to indicate which CacheDirective you want to modify.
- @param flags {@link CacheFlag}s to use for this operation.
- @throws IOException if the directive could not be modified]]>
- </doc>
- </method>
- <method name="removeCacheDirective"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <param name="id" type="long"/>
- <exception name="IOException" type="java.io.IOException"/>
- <doc>
- <![CDATA[Remove a CacheDirective.
- @param id identifier of the CacheDirectiveInfo to remove
- @throws IOException if the directive could not be removed]]>
- </doc>
- </method>
- <method name="listCacheDirectives" return="org.apache.hadoop.fs.RemoteIterator"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <param name="filter" type="org.apache.hadoop.hdfs.protocol.CacheDirectiveInfo"/>
- <exception name="IOException" type="java.io.IOException"/>
- <doc>
- <![CDATA[List cache directives. Incrementally fetches results from the server.
- @param filter Filter parameters to use when listing the directives, null to
- list all directives visible to us.
- @return A RemoteIterator which returns CacheDirectiveInfo objects.]]>
- </doc>
- </method>
- <method name="addCachePool"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <param name="info" type="org.apache.hadoop.hdfs.protocol.CachePoolInfo"/>
- <exception name="IOException" type="java.io.IOException"/>
- <doc>
- <![CDATA[Add a cache pool.
- @param info
- The request to add a cache pool.
- @throws IOException
- If the request could not be completed.]]>
- </doc>
- </method>
- <method name="modifyCachePool"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <param name="info" type="org.apache.hadoop.hdfs.protocol.CachePoolInfo"/>
- <exception name="IOException" type="java.io.IOException"/>
- <doc>
- <![CDATA[Modify an existing cache pool.
- @param info
- The request to modify a cache pool.
- @throws IOException
- If the request could not be completed.]]>
- </doc>
- </method>
- <method name="removeCachePool"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <param name="poolName" type="java.lang.String"/>
- <exception name="IOException" type="java.io.IOException"/>
- <doc>
- <![CDATA[Remove a cache pool.
- @param poolName
- Name of the cache pool to remove.
- @throws IOException
- if the cache pool did not exist, or could not be removed.]]>
- </doc>
- </method>
- <method name="listCachePools" return="org.apache.hadoop.fs.RemoteIterator"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <exception name="IOException" type="java.io.IOException"/>
- <doc>
- <![CDATA[List all cache pools.
- @return A remote iterator from which you can get CachePoolEntry objects.
- Requests will be made as needed.
- @throws IOException
- If there was an error listing cache pools.]]>
- </doc>
- </method>
- <method name="createEncryptionZone"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <param name="path" type="org.apache.hadoop.fs.Path"/>
- <param name="keyName" type="java.lang.String"/>
- <exception name="IOException" type="java.io.IOException"/>
- <exception name="AccessControlException" type="org.apache.hadoop.security.AccessControlException"/>
- <exception name="FileNotFoundException" type="java.io.FileNotFoundException"/>
- <doc>
- <![CDATA[Create an encryption zone rooted at an empty existing directory, using the
- specified encryption key. An encryption zone has an associated encryption
- key used when reading and writing files within the zone.
- @param path The path of the root of the encryption zone. Must refer to
- an empty, existing directory.
- @param keyName Name of key available at the KeyProvider.
- @throws IOException if there was a general IO exception
- @throws AccessControlException if the caller does not have access to path
- @throws FileNotFoundException if the path does not exist]]>
- </doc>
- </method>
- <method name="getEncryptionZoneForPath" return="org.apache.hadoop.hdfs.protocol.EncryptionZone"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <param name="path" type="org.apache.hadoop.fs.Path"/>
- <exception name="IOException" type="java.io.IOException"/>
- <exception name="AccessControlException" type="org.apache.hadoop.security.AccessControlException"/>
- <exception name="FileNotFoundException" type="java.io.FileNotFoundException"/>
- <doc>
- <![CDATA[Get the path of the encryption zone for a given file or directory.
- @param path The path to get the ez for.
- @return The EncryptionZone of the ez, or null if path is not in an ez.
- @throws IOException if there was a general IO exception
- @throws AccessControlException if the caller does not have access to path
- @throws FileNotFoundException if the path does not exist]]>
- </doc>
- </method>
- <method name="listEncryptionZones" return="org.apache.hadoop.fs.RemoteIterator"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <exception name="IOException" type="java.io.IOException"/>
- <doc>
- <![CDATA[Returns a RemoteIterator which can be used to list the encryption zones
- in HDFS. For large numbers of encryption zones, the iterator will fetch
- the list of zones in a number of small batches.
- <p/>
- Since the list is fetched in batches, it does not represent a
- consistent snapshot of the entire list of encryption zones.
- <p/>
- This method can only be called by HDFS superusers.]]>
- </doc>
- </method>
- <method name="getInotifyEventStream" return="org.apache.hadoop.hdfs.DFSInotifyEventInputStream"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <exception name="IOException" type="java.io.IOException"/>
- <doc>
- <![CDATA[Exposes a stream of namesystem events. Only events occurring after the
- stream is created are available.
- See {@link org.apache.hadoop.hdfs.DFSInotifyEventInputStream}
- for information on stream usage.
- See {@link org.apache.hadoop.hdfs.inotify.Event}
- for information on the available events.
- <p/>
- Inotify users may want to tune the following HDFS parameters to
- ensure that enough extra HDFS edits are saved to support inotify clients
- that fall behind the current state of the namespace while reading events.
- The default parameter values should generally be reasonable. If edits are
- deleted before their corresponding events can be read, clients will see a
- {@link org.apache.hadoop.hdfs.inotify.MissingEventsException} on
- {@link org.apache.hadoop.hdfs.DFSInotifyEventInputStream} method calls.
- It should generally be sufficient to tune these parameters:
- dfs.namenode.num.extra.edits.retained
- dfs.namenode.max.extra.edits.segments.retained
- Parameters that affect the number of created segments and the number of
- edits that are considered necessary, i.e. do not count towards the
- dfs.namenode.num.extra.edits.retained quota):
- dfs.namenode.checkpoint.period
- dfs.namenode.checkpoint.txns
- dfs.namenode.num.checkpoints.retained
- dfs.ha.log-roll.period
- <p/>
- It is recommended that local journaling be configured
- (dfs.namenode.edits.dir) for inotify (in addition to a shared journal)
- so that edit transfers from the shared journal can be avoided.
- @throws IOException If there was an error obtaining the stream.]]>
- </doc>
- </method>
- <method name="getInotifyEventStream" return="org.apache.hadoop.hdfs.DFSInotifyEventInputStream"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <param name="lastReadTxid" type="long"/>
- <exception name="IOException" type="java.io.IOException"/>
- <doc>
- <![CDATA[A version of {@link HdfsAdmin#getInotifyEventStream()} meant for advanced
- users who are aware of HDFS edits up to lastReadTxid (e.g. because they
- have access to an FSImage inclusive of lastReadTxid) and only want to read
- events after this point.]]>
- </doc>
- </method>
- <method name="setStoragePolicy"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <param name="src" type="org.apache.hadoop.fs.Path"/>
- <param name="policyName" type="java.lang.String"/>
- <exception name="IOException" type="java.io.IOException"/>
- <doc>
- <![CDATA[Set the source path to the specified storage policy.
- @param src The source path referring to either a directory or a file.
- @param policyName The name of the storage policy.]]>
- </doc>
- </method>
- <doc>
- <![CDATA[The public API for performing administrative functions on HDFS. Those writing
- applications against HDFS should prefer this interface to directly accessing
- functionality in DistributedFileSystem or DFSClient.
- Note that this is distinct from the similarly-named {@link DFSAdmin}, which
- is a class that provides the functionality for the CLI `hdfs dfsadmin ...'
- commands.]]>
- </doc>
- </class>
- <!-- end class org.apache.hadoop.hdfs.client.HdfsAdmin -->
- <!-- start class org.apache.hadoop.hdfs.client.HdfsDataInputStream -->
- <class name="HdfsDataInputStream" extends="org.apache.hadoop.fs.FSDataInputStream"
- abstract="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <constructor name="HdfsDataInputStream" type="org.apache.hadoop.hdfs.DFSInputStream"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <exception name="IOException" type="java.io.IOException"/>
- </constructor>
- <constructor name="HdfsDataInputStream" type="org.apache.hadoop.crypto.CryptoInputStream"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <exception name="IOException" type="java.io.IOException"/>
- </constructor>
- <method name="getWrappedStream" return="java.io.InputStream"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[Get a reference to the wrapped output stream. We always want to return the
- actual underlying InputStream, even when we're using a CryptoStream. e.g.
- in the delegated methods below.
- @return the underlying output stream]]>
- </doc>
- </method>
- <method name="getCurrentDatanode" return="org.apache.hadoop.hdfs.protocol.DatanodeInfo"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[Get the datanode from which the stream is currently reading.]]>
- </doc>
- </method>
- <method name="getCurrentBlock" return="org.apache.hadoop.hdfs.protocol.ExtendedBlock"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[Get the block containing the target position.]]>
- </doc>
- </method>
- <method name="getAllBlocks" return="java.util.List"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <exception name="IOException" type="java.io.IOException"/>
- <doc>
- <![CDATA[Get the collection of blocks that has already been located.]]>
- </doc>
- </method>
- <method name="getVisibleLength" return="long"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <exception name="IOException" type="java.io.IOException"/>
- <doc>
- <![CDATA[Get the visible length of the file. It will include the length of the last
- block even if that is in UnderConstruction state.
- @return The visible length of the file.]]>
- </doc>
- </method>
- <method name="getReadStatistics" return="org.apache.hadoop.hdfs.DFSInputStream.ReadStatistics"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[Get statistics about the reads which this DFSInputStream has done.
- Note that because HdfsDataInputStream is buffered, these stats may
- be higher than you would expect just by adding up the number of
- bytes read through HdfsDataInputStream.]]>
- </doc>
- </method>
- <method name="clearReadStatistics"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- </method>
- <doc>
- <![CDATA[The Hdfs implementation of {@link FSDataInputStream}.]]>
- </doc>
- </class>
- <!-- end class org.apache.hadoop.hdfs.client.HdfsDataInputStream -->
- <!-- start class org.apache.hadoop.hdfs.client.HdfsDataOutputStream -->
- <class name="HdfsDataOutputStream" extends="org.apache.hadoop.fs.FSDataOutputStream"
- abstract="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <constructor name="HdfsDataOutputStream" type="org.apache.hadoop.hdfs.DFSOutputStream, org.apache.hadoop.fs.FileSystem.Statistics, long"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <exception name="IOException" type="java.io.IOException"/>
- </constructor>
- <constructor name="HdfsDataOutputStream" type="org.apache.hadoop.hdfs.DFSOutputStream, org.apache.hadoop.fs.FileSystem.Statistics"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <exception name="IOException" type="java.io.IOException"/>
- </constructor>
- <constructor name="HdfsDataOutputStream" type="org.apache.hadoop.crypto.CryptoOutputStream, org.apache.hadoop.fs.FileSystem.Statistics, long"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <exception name="IOException" type="java.io.IOException"/>
- </constructor>
- <constructor name="HdfsDataOutputStream" type="org.apache.hadoop.crypto.CryptoOutputStream, org.apache.hadoop.fs.FileSystem.Statistics"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <exception name="IOException" type="java.io.IOException"/>
- </constructor>
- <method name="getCurrentBlockReplication" return="int"
- abstract="false" native="false" synchronized="true"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <exception name="IOException" type="java.io.IOException"/>
- <doc>
- <![CDATA[Get the actual number of replicas of the current block.
- This can be different from the designated replication factor of the file
- because the namenode does not maintain replication for the blocks which are
- currently being written to. Depending on the configuration, the client may
- continue to write to a block even if a few datanodes in the write pipeline
- have failed, or the client may add a new datanodes once a datanode has
- failed.
- @return the number of valid replicas of the current block]]>
- </doc>
- </method>
- <method name="hsync"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <param name="syncFlags" type="java.util.EnumSet"/>
- <exception name="IOException" type="java.io.IOException"/>
- <doc>
- <![CDATA[Sync buffered data to DataNodes (flush to disk devices).
- @param syncFlags
- Indicate the detailed semantic and actions of the hsync.
- @throws IOException
- @see FSDataOutputStream#hsync()]]>
- </doc>
- </method>
- <doc>
- <![CDATA[The Hdfs implementation of {@link FSDataOutputStream}.]]>
- </doc>
- </class>
- <!-- end class org.apache.hadoop.hdfs.client.HdfsDataOutputStream -->
- <!-- start class org.apache.hadoop.hdfs.client.HdfsUtils -->
- <class name="HdfsUtils" extends="java.lang.Object"
- abstract="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <constructor name="HdfsUtils"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- </constructor>
- <method name="isHealthy" return="boolean"
- abstract="false" native="false" synchronized="false"
- static="true" final="false" visibility="public"
- deprecated="not deprecated">
- <param name="uri" type="java.net.URI"/>
- <doc>
- <![CDATA[Is the HDFS healthy?
- HDFS is considered as healthy if it is up and not in safemode.
- @param uri the HDFS URI. Note that the URI path is ignored.
- @return true if HDFS is healthy; false, otherwise.]]>
- </doc>
- </method>
- <doc>
- <![CDATA[The public utility API for HDFS.]]>
- </doc>
- </class>
- <!-- end class org.apache.hadoop.hdfs.client.HdfsUtils -->
- </package>
- <package name="org.apache.hadoop.hdfs.inotify">
- <!-- start class org.apache.hadoop.hdfs.inotify.Event -->
- <class name="Event" extends="java.lang.Object"
- abstract="true"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <constructor name="Event" type="org.apache.hadoop.hdfs.inotify.Event.EventType"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- </constructor>
- <method name="getEventType" return="org.apache.hadoop.hdfs.inotify.Event.EventType"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- </method>
- <doc>
- <![CDATA[Events sent by the inotify system. Note that no events are necessarily sent
- when a file is opened for read (although a MetadataUpdateEvent will be sent
- if the atime is updated).]]>
- </doc>
- </class>
- <!-- end class org.apache.hadoop.hdfs.inotify.Event -->
- <!-- start class org.apache.hadoop.hdfs.inotify.EventBatch -->
- <class name="EventBatch" extends="java.lang.Object"
- abstract="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <constructor name="EventBatch" type="long, org.apache.hadoop.hdfs.inotify.Event[]"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- </constructor>
- <method name="getTxid" return="long"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- </method>
- <method name="getEvents" return="org.apache.hadoop.hdfs.inotify.Event[]"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- </method>
- <doc>
- <![CDATA[A batch of events that all happened on the same transaction ID.]]>
- </doc>
- </class>
- <!-- end class org.apache.hadoop.hdfs.inotify.EventBatch -->
- <!-- start class org.apache.hadoop.hdfs.inotify.MissingEventsException -->
- <class name="MissingEventsException" extends="java.lang.Exception"
- abstract="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <constructor name="MissingEventsException"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- </constructor>
- <constructor name="MissingEventsException" type="long, long"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- </constructor>
- <method name="getExpectedTxid" return="long"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- </method>
- <method name="getActualTxid" return="long"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- </method>
- <method name="toString" return="java.lang.String"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- </method>
- </class>
- <!-- end class org.apache.hadoop.hdfs.inotify.MissingEventsException -->
- </package>
- <package name="org.apache.hadoop.hdfs.net">
- </package>
- <package name="org.apache.hadoop.hdfs.protocol">
- <!-- start class org.apache.hadoop.hdfs.protocol.CacheDirectiveEntry -->
- <class name="CacheDirectiveEntry" extends="java.lang.Object"
- abstract="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <constructor name="CacheDirectiveEntry" type="org.apache.hadoop.hdfs.protocol.CacheDirectiveInfo, org.apache.hadoop.hdfs.protocol.CacheDirectiveStats"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- </constructor>
- <method name="getInfo" return="org.apache.hadoop.hdfs.protocol.CacheDirectiveInfo"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- </method>
- <method name="getStats" return="org.apache.hadoop.hdfs.protocol.CacheDirectiveStats"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- </method>
- <doc>
- <![CDATA[Describes a path-based cache directive entry.]]>
- </doc>
- </class>
- <!-- end class org.apache.hadoop.hdfs.protocol.CacheDirectiveEntry -->
- <!-- start class org.apache.hadoop.hdfs.protocol.CacheDirectiveInfo -->
- <class name="CacheDirectiveInfo" extends="java.lang.Object"
- abstract="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <method name="getId" return="java.lang.Long"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[@return The ID of this directive.]]>
- </doc>
- </method>
- <method name="getPath" return="org.apache.hadoop.fs.Path"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[@return The path used in this request.]]>
- </doc>
- </method>
- <method name="getReplication" return="java.lang.Short"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[@return The number of times the block should be cached.]]>
- </doc>
- </method>
- <method name="getPool" return="java.lang.String"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[@return The pool used in this request.]]>
- </doc>
- </method>
- <method name="getExpiration" return="org.apache.hadoop.hdfs.protocol.CacheDirectiveInfo.Expiration"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[@return When this directive expires.]]>
- </doc>
- </method>
- <method name="equals" return="boolean"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <param name="o" type="java.lang.Object"/>
- </method>
- <method name="hashCode" return="int"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- </method>
- <method name="toString" return="java.lang.String"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- </method>
- <doc>
- <![CDATA[Describes a path-based cache directive.]]>
- </doc>
- </class>
- <!-- end class org.apache.hadoop.hdfs.protocol.CacheDirectiveInfo -->
- <!-- start class org.apache.hadoop.hdfs.protocol.CacheDirectiveStats -->
- <class name="CacheDirectiveStats" extends="java.lang.Object"
- abstract="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <method name="getBytesNeeded" return="long"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[@return The bytes needed.]]>
- </doc>
- </method>
- <method name="getBytesCached" return="long"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[@return The bytes cached.]]>
- </doc>
- </method>
- <method name="getFilesNeeded" return="long"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[@return The number of files needed.]]>
- </doc>
- </method>
- <method name="getFilesCached" return="long"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[@return The number of files cached.]]>
- </doc>
- </method>
- <method name="hasExpired" return="boolean"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[@return Whether this directive has expired.]]>
- </doc>
- </method>
- <method name="toString" return="java.lang.String"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- </method>
- <doc>
- <![CDATA[Describes a path-based cache directive.]]>
- </doc>
- </class>
- <!-- end class org.apache.hadoop.hdfs.protocol.CacheDirectiveStats -->
- <!-- start class org.apache.hadoop.hdfs.protocol.CachePoolEntry -->
- <class name="CachePoolEntry" extends="java.lang.Object"
- abstract="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <constructor name="CachePoolEntry" type="org.apache.hadoop.hdfs.protocol.CachePoolInfo, org.apache.hadoop.hdfs.protocol.CachePoolStats"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- </constructor>
- <method name="getInfo" return="org.apache.hadoop.hdfs.protocol.CachePoolInfo"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- </method>
- <method name="getStats" return="org.apache.hadoop.hdfs.protocol.CachePoolStats"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- </method>
- <doc>
- <![CDATA[Describes a Cache Pool entry.]]>
- </doc>
- </class>
- <!-- end class org.apache.hadoop.hdfs.protocol.CachePoolEntry -->
- <!-- start class org.apache.hadoop.hdfs.protocol.CachePoolInfo -->
- <class name="CachePoolInfo" extends="java.lang.Object"
- abstract="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <constructor name="CachePoolInfo" type="java.lang.String"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- </constructor>
- <method name="getPoolName" return="java.lang.String"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[@return Name of the pool.]]>
- </doc>
- </method>
- <method name="getOwnerName" return="java.lang.String"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[@return The owner of the pool. Along with the group and mode, determines
- who has access to view and modify the pool.]]>
- </doc>
- </method>
- <method name="setOwnerName" return="org.apache.hadoop.hdfs.protocol.CachePoolInfo"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <param name="ownerName" type="java.lang.String"/>
- </method>
- <method name="getGroupName" return="java.lang.String"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[@return The group of the pool. Along with the owner and mode, determines
- who has access to view and modify the pool.]]>
- </doc>
- </method>
- <method name="setGroupName" return="org.apache.hadoop.hdfs.protocol.CachePoolInfo"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <param name="groupName" type="java.lang.String"/>
- </method>
- <method name="getMode" return="org.apache.hadoop.fs.permission.FsPermission"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[@return Unix-style permissions of the pool. Along with the owner and group,
- determines who has access to view and modify the pool.]]>
- </doc>
- </method>
- <method name="setMode" return="org.apache.hadoop.hdfs.protocol.CachePoolInfo"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <param name="mode" type="org.apache.hadoop.fs.permission.FsPermission"/>
- </method>
- <method name="getLimit" return="java.lang.Long"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[@return The maximum aggregate number of bytes that can be cached by
- directives in this pool.]]>
- </doc>
- </method>
- <method name="setLimit" return="org.apache.hadoop.hdfs.protocol.CachePoolInfo"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <param name="bytes" type="java.lang.Long"/>
- </method>
- <method name="getMaxRelativeExpiryMs" return="java.lang.Long"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[@return The maximum relative expiration of directives of this pool in
- milliseconds]]>
- </doc>
- </method>
- <method name="setMaxRelativeExpiryMs" return="org.apache.hadoop.hdfs.protocol.CachePoolInfo"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <param name="ms" type="java.lang.Long"/>
- <doc>
- <![CDATA[Set the maximum relative expiration of directives of this pool in
- milliseconds.
- @param ms in milliseconds
- @return This builder, for call chaining.]]>
- </doc>
- </method>
- <method name="toString" return="java.lang.String"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- </method>
- <method name="equals" return="boolean"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <param name="o" type="java.lang.Object"/>
- </method>
- <method name="hashCode" return="int"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- </method>
- <method name="validate"
- abstract="false" native="false" synchronized="false"
- static="true" final="false" visibility="public"
- deprecated="not deprecated">
- <param name="info" type="org.apache.hadoop.hdfs.protocol.CachePoolInfo"/>
- <exception name="IOException" type="java.io.IOException"/>
- </method>
- <method name="validateName"
- abstract="false" native="false" synchronized="false"
- static="true" final="false" visibility="public"
- deprecated="not deprecated">
- <param name="poolName" type="java.lang.String"/>
- <exception name="IOException" type="java.io.IOException"/>
- </method>
- <field name="LOG" type="org.apache.commons.logging.Log"
- transient="false" volatile="false"
- static="true" final="true" visibility="public"
- deprecated="not deprecated">
- </field>
- <field name="RELATIVE_EXPIRY_NEVER" type="long"
- transient="false" volatile="false"
- static="true" final="true" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[Indicates that the pool does not have a maximum relative expiry.]]>
- </doc>
- </field>
- <field name="DEFAULT_MAX_RELATIVE_EXPIRY" type="long"
- transient="false" volatile="false"
- static="true" final="true" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[Default max relative expiry for cache pools.]]>
- </doc>
- </field>
- <field name="LIMIT_UNLIMITED" type="long"
- transient="false" volatile="false"
- static="true" final="true" visibility="public"
- deprecated="not deprecated">
- </field>
- <field name="DEFAULT_LIMIT" type="long"
- transient="false" volatile="false"
- static="true" final="true" visibility="public"
- deprecated="not deprecated">
- </field>
- <doc>
- <![CDATA[CachePoolInfo describes a cache pool.
- This class is used in RPCs to create and modify cache pools.
- It is serializable and can be stored in the edit log.]]>
- </doc>
- </class>
- <!-- end class org.apache.hadoop.hdfs.protocol.CachePoolInfo -->
- <!-- start class org.apache.hadoop.hdfs.protocol.CachePoolStats -->
- <class name="CachePoolStats" extends="java.lang.Object"
- abstract="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <method name="getBytesNeeded" return="long"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- </method>
- <method name="getBytesCached" return="long"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- </method>
- <method name="getBytesOverlimit" return="long"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- </method>
- <method name="getFilesNeeded" return="long"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- </method>
- <method name="getFilesCached" return="long"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- </method>
- <method name="toString" return="java.lang.String"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- </method>
- <doc>
- <![CDATA[CachePoolStats describes cache pool statistics.]]>
- </doc>
- </class>
- <!-- end class org.apache.hadoop.hdfs.protocol.CachePoolStats -->
- <!-- start class org.apache.hadoop.hdfs.protocol.EncryptionZone -->
- <class name="EncryptionZone" extends="java.lang.Object"
- abstract="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <constructor name="EncryptionZone" type="long, java.lang.String, org.apache.hadoop.crypto.CipherSuite, org.apache.hadoop.crypto.CryptoProtocolVersion, java.lang.String"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- </constructor>
- <method name="getId" return="long"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- </method>
- <method name="getPath" return="java.lang.String"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- </method>
- <method name="getSuite" return="org.apache.hadoop.crypto.CipherSuite"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- </method>
- <method name="getVersion" return="org.apache.hadoop.crypto.CryptoProtocolVersion"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- </method>
- <method name="getKeyName" return="java.lang.String"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- </method>
- <method name="hashCode" return="int"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- </method>
- <method name="equals" return="boolean"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <param name="obj" type="java.lang.Object"/>
- </method>
- <method name="toString" return="java.lang.String"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- </method>
- <doc>
- <![CDATA[A simple class for representing an encryption zone. Presently an encryption
- zone only has a path (the root of the encryption zone), a key name, and a
- unique id. The id is used to implement batched listing of encryption zones.]]>
- </doc>
- </class>
- <!-- end class org.apache.hadoop.hdfs.protocol.EncryptionZone -->
- </package>
- <package name="org.apache.hadoop.hdfs.protocol.datatransfer">
- </package>
- <package name="org.apache.hadoop.hdfs.protocol.datatransfer.sasl">
- </package>
- <package name="org.apache.hadoop.hdfs.protocolPB">
- </package>
- <package name="org.apache.hadoop.hdfs.qjournal.client">
- </package>
- <package name="org.apache.hadoop.hdfs.qjournal.protocol">
- </package>
- <package name="org.apache.hadoop.hdfs.qjournal.protocolPB">
- </package>
- <package name="org.apache.hadoop.hdfs.qjournal.server">
- <!-- start interface org.apache.hadoop.hdfs.qjournal.server.JournalNodeMXBean -->
- <interface name="JournalNodeMXBean" abstract="true"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <method name="getJournalsStatus" return="java.lang.String"
- abstract="true" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[Get status information (e.g., whether formatted) of JournalNode's journals.
- @return A string presenting status for each journal]]>
- </doc>
- </method>
- <doc>
- <![CDATA[This is the JMX management interface for JournalNode information]]>
- </doc>
- </interface>
- <!-- end interface org.apache.hadoop.hdfs.qjournal.server.JournalNodeMXBean -->
- </package>
- <package name="org.apache.hadoop.hdfs.security.token.block">
- </package>
- <package name="org.apache.hadoop.hdfs.security.token.delegation">
- </package>
- <package name="org.apache.hadoop.hdfs.server.balancer">
- </package>
- <package name="org.apache.hadoop.hdfs.server.blockmanagement">
- </package>
- <package name="org.apache.hadoop.hdfs.server.common">
- </package>
- <package name="org.apache.hadoop.hdfs.server.datanode">
- <!-- start interface org.apache.hadoop.hdfs.server.datanode.DataNodeMXBean -->
- <interface name="DataNodeMXBean" abstract="true"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <method name="getVersion" return="java.lang.String"
- abstract="true" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[Gets the version of Hadoop.
- @return the version of Hadoop]]>
- </doc>
- </method>
- <method name="getRpcPort" return="java.lang.String"
- abstract="true" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[Gets the rpc port.
- @return the rpc port]]>
- </doc>
- </method>
- <method name="getHttpPort" return="java.lang.String"
- abstract="true" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[Gets the http port.
- @return the http port]]>
- </doc>
- </method>
- <method name="getNamenodeAddresses" return="java.lang.String"
- abstract="true" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[Gets the namenode IP addresses
- @return the namenode IP addresses that the datanode is talking to]]>
- </doc>
- </method>
- <method name="getVolumeInfo" return="java.lang.String"
- abstract="true" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[Gets the information of each volume on the Datanode. Please
- see the implementation for the format of returned information.
- @return the volume info]]>
- </doc>
- </method>
- <method name="getClusterId" return="java.lang.String"
- abstract="true" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[Gets the cluster id.
- @return the cluster id]]>
- </doc>
- </method>
- <method name="getXceiverCount" return="int"
- abstract="true" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[Returns an estimate of the number of Datanode threads
- actively transferring blocks.]]>
- </doc>
- </method>
- <method name="getDatanodeNetworkCounts" return="java.util.Map"
- abstract="true" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[Gets the network error counts on a per-Datanode basis.]]>
- </doc>
- </method>
- <doc>
- <![CDATA[This is the JMX management interface for data node information]]>
- </doc>
- </interface>
- <!-- end interface org.apache.hadoop.hdfs.server.datanode.DataNodeMXBean -->
- </package>
- <package name="org.apache.hadoop.hdfs.server.datanode.fsdataset">
- </package>
- <package name="org.apache.hadoop.hdfs.server.datanode.fsdataset.impl">
- </package>
- <package name="org.apache.hadoop.hdfs.server.datanode.metrics">
- </package>
- <package name="org.apache.hadoop.hdfs.server.datanode.web">
- </package>
- <package name="org.apache.hadoop.hdfs.server.datanode.web.webhdfs">
- </package>
- <package name="org.apache.hadoop.hdfs.server.mover">
- </package>
- <package name="org.apache.hadoop.hdfs.server.namenode">
- <!-- start interface org.apache.hadoop.hdfs.server.namenode.AuditLogger -->
- <interface name="AuditLogger" abstract="true"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <method name="initialize"
- abstract="true" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <param name="conf" type="org.apache.hadoop.conf.Configuration"/>
- <doc>
- <![CDATA[Called during initialization of the logger.
- @param conf The configuration object.]]>
- </doc>
- </method>
- <method name="logAuditEvent"
- abstract="true" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <param name="succeeded" type="boolean"/>
- <param name="userName" type="java.lang.String"/>
- <param name="addr" type="java.net.InetAddress"/>
- <param name="cmd" type="java.lang.String"/>
- <param name="src" type="java.lang.String"/>
- <param name="dst" type="java.lang.String"/>
- <param name="stat" type="org.apache.hadoop.fs.FileStatus"/>
- <doc>
- <![CDATA[Called to log an audit event.
- <p>
- This method must return as quickly as possible, since it's called
- in a critical section of the NameNode's operation.
- @param succeeded Whether authorization succeeded.
- @param userName Name of the user executing the request.
- @param addr Remote address of the request.
- @param cmd The requested command.
- @param src Path of affected source file.
- @param dst Path of affected destination file (if any).
- @param stat File information for operations that change the file's
- metadata (permissions, owner, times, etc).]]>
- </doc>
- </method>
- <doc>
- <![CDATA[Interface defining an audit logger.]]>
- </doc>
- </interface>
- <!-- end interface org.apache.hadoop.hdfs.server.namenode.AuditLogger -->
- <!-- start class org.apache.hadoop.hdfs.server.namenode.HdfsAuditLogger -->
- <class name="HdfsAuditLogger" extends="java.lang.Object"
- abstract="true"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <implements name="org.apache.hadoop.hdfs.server.namenode.AuditLogger"/>
- <constructor name="HdfsAuditLogger"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- </constructor>
- <method name="logAuditEvent"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <param name="succeeded" type="boolean"/>
- <param name="userName" type="java.lang.String"/>
- <param name="addr" type="java.net.InetAddress"/>
- <param name="cmd" type="java.lang.String"/>
- <param name="src" type="java.lang.String"/>
- <param name="dst" type="java.lang.String"/>
- <param name="status" type="org.apache.hadoop.fs.FileStatus"/>
- </method>
- <method name="logAuditEvent"
- abstract="true" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <param name="succeeded" type="boolean"/>
- <param name="userName" type="java.lang.String"/>
- <param name="addr" type="java.net.InetAddress"/>
- <param name="cmd" type="java.lang.String"/>
- <param name="src" type="java.lang.String"/>
- <param name="dst" type="java.lang.String"/>
- <param name="stat" type="org.apache.hadoop.fs.FileStatus"/>
- <param name="ugi" type="org.apache.hadoop.security.UserGroupInformation"/>
- <param name="dtSecretManager" type="org.apache.hadoop.hdfs.security.token.delegation.DelegationTokenSecretManager"/>
- <doc>
- <![CDATA[Same as
- {@link #logAuditEvent(boolean, String, InetAddress, String, String, String, FileStatus)}
- with additional parameters related to logging delegation token tracking
- IDs.
- @param succeeded Whether authorization succeeded.
- @param userName Name of the user executing the request.
- @param addr Remote address of the request.
- @param cmd The requested command.
- @param src Path of affected source file.
- @param dst Path of affected destination file (if any).
- @param stat File information for operations that change the file's metadata
- (permissions, owner, times, etc).
- @param ugi UserGroupInformation of the current user, or null if not logging
- token tracking information
- @param dtSecretManager The token secret manager, or null if not logging
- token tracking information]]>
- </doc>
- </method>
- <doc>
- <![CDATA[Extension of {@link AuditLogger}.]]>
- </doc>
- </class>
- <!-- end class org.apache.hadoop.hdfs.server.namenode.HdfsAuditLogger -->
- <!-- start class org.apache.hadoop.hdfs.server.namenode.INodeAttributeProvider -->
- <class name="INodeAttributeProvider" extends="java.lang.Object"
- abstract="true"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <constructor name="INodeAttributeProvider"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- </constructor>
- <method name="start"
- abstract="true" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[Initialize the provider. This method is called at NameNode startup
- time.]]>
- </doc>
- </method>
- <method name="stop"
- abstract="true" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[Shutdown the provider. This method is called at NameNode shutdown time.]]>
- </doc>
- </method>
- <method name="getAttributes" return="org.apache.hadoop.hdfs.server.namenode.INodeAttributes"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <param name="fullPath" type="java.lang.String"/>
- <param name="inode" type="org.apache.hadoop.hdfs.server.namenode.INodeAttributes"/>
- </method>
- <method name="getAttributes" return="org.apache.hadoop.hdfs.server.namenode.INodeAttributes"
- abstract="true" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <param name="pathElements" type="java.lang.String[]"/>
- <param name="inode" type="org.apache.hadoop.hdfs.server.namenode.INodeAttributes"/>
- </method>
- <method name="getExternalAccessControlEnforcer" return="org.apache.hadoop.hdfs.server.namenode.INodeAttributeProvider.AccessControlEnforcer"
- abstract="false" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <param name="defaultEnforcer" type="org.apache.hadoop.hdfs.server.namenode.INodeAttributeProvider.AccessControlEnforcer"/>
- <doc>
- <![CDATA[Can be over-ridden by implementations to provide a custom Access Control
- Enforcer that can provide an alternate implementation of the
- default permission checking logic.
- @param defaultEnforcer The Default AccessControlEnforcer
- @return The AccessControlEnforcer to use]]>
- </doc>
- </method>
- </class>
- <!-- end class org.apache.hadoop.hdfs.server.namenode.INodeAttributeProvider -->
- <!-- start interface org.apache.hadoop.hdfs.server.namenode.NameNodeMXBean -->
- <interface name="NameNodeMXBean" abstract="true"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <method name="getVersion" return="java.lang.String"
- abstract="true" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[Gets the version of Hadoop.
- @return the version]]>
- </doc>
- </method>
- <method name="getSoftwareVersion" return="java.lang.String"
- abstract="true" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[Get the version of software running on the Namenode
- @return a string representing the version]]>
- </doc>
- </method>
- <method name="getUsed" return="long"
- abstract="true" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[Gets the used space by data nodes.
- @return the used space by data nodes]]>
- </doc>
- </method>
- <method name="getFree" return="long"
- abstract="true" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[Gets total non-used raw bytes.
- @return total non-used raw bytes]]>
- </doc>
- </method>
- <method name="getTotal" return="long"
- abstract="true" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[Gets total raw bytes including non-dfs used space.
- @return the total raw bytes including non-dfs used space]]>
- </doc>
- </method>
- <method name="getSafemode" return="java.lang.String"
- abstract="true" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[Gets the safemode status
- @return the safemode status]]>
- </doc>
- </method>
- <method name="isUpgradeFinalized" return="boolean"
- abstract="true" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[Checks if upgrade is finalized.
- @return true, if upgrade is finalized]]>
- </doc>
- </method>
- <method name="getRollingUpgradeStatus" return="org.apache.hadoop.hdfs.protocol.RollingUpgradeInfo.Bean"
- abstract="true" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[Gets the RollingUpgrade information.
- @return Rolling upgrade information if an upgrade is in progress. Else
- (e.g. if there is no upgrade or the upgrade is finalized), returns null.]]>
- </doc>
- </method>
- <method name="getNonDfsUsedSpace" return="long"
- abstract="true" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[Gets total used space by data nodes for non DFS purposes such as storing
- temporary files on the local file system
- @return the non dfs space of the cluster]]>
- </doc>
- </method>
- <method name="getPercentUsed" return="float"
- abstract="true" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[Gets the total used space by data nodes as percentage of total capacity
- @return the percentage of used space on the cluster.]]>
- </doc>
- </method>
- <method name="getPercentRemaining" return="float"
- abstract="true" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[Gets the total remaining space by data nodes as percentage of total
- capacity
- @return the percentage of the remaining space on the cluster]]>
- </doc>
- </method>
- <method name="getCacheUsed" return="long"
- abstract="true" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[Returns the amount of cache used by the datanode (in bytes).]]>
- </doc>
- </method>
- <method name="getCacheCapacity" return="long"
- abstract="true" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[Returns the total cache capacity of the datanode (in bytes).]]>
- </doc>
- </method>
- <method name="getBlockPoolUsedSpace" return="long"
- abstract="true" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[Get the total space used by the block pools of this namenode]]>
- </doc>
- </method>
- <method name="getPercentBlockPoolUsed" return="float"
- abstract="true" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[Get the total space used by the block pool as percentage of total capacity]]>
- </doc>
- </method>
- <method name="getTotalBlocks" return="long"
- abstract="true" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[Gets the total numbers of blocks on the cluster.
- @return the total number of blocks of the cluster]]>
- </doc>
- </method>
- <method name="getTotalFiles" return="long"
- abstract="true" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[Gets the total number of files on the cluster
- @return the total number of files on the cluster]]>
- </doc>
- </method>
- <method name="getNumberOfMissingBlocks" return="long"
- abstract="true" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[Gets the total number of missing blocks on the cluster
- @return the total number of missing blocks on the cluster]]>
- </doc>
- </method>
- <method name="getNumberOfMissingBlocksWithReplicationFactorOne" return="long"
- abstract="true" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[Gets the total number of missing blocks on the cluster with
- replication factor 1
- @return the total number of missing blocks on the cluster with
- replication factor 1]]>
- </doc>
- </method>
- <method name="getThreads" return="int"
- abstract="true" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[Gets the number of threads.
- @return the number of threads]]>
- </doc>
- </method>
- <method name="getLiveNodes" return="java.lang.String"
- abstract="true" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[Gets the live node information of the cluster.
- @return the live node information]]>
- </doc>
- </method>
- <method name="getDeadNodes" return="java.lang.String"
- abstract="true" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[Gets the dead node information of the cluster.
- @return the dead node information]]>
- </doc>
- </method>
- <method name="getDecomNodes" return="java.lang.String"
- abstract="true" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[Gets the decommissioning node information of the cluster.
- @return the decommissioning node information]]>
- </doc>
- </method>
- <method name="getClusterId" return="java.lang.String"
- abstract="true" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[Gets the cluster id.
- @return the cluster id]]>
- </doc>
- </method>
- <method name="getBlockPoolId" return="java.lang.String"
- abstract="true" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[Gets the block pool id.
- @return the block pool id]]>
- </doc>
- </method>
- <method name="getNameDirStatuses" return="java.lang.String"
- abstract="true" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[Get status information about the directories storing image and edits logs
- of the NN.
- @return the name dir status information, as a JSON string.]]>
- </doc>
- </method>
- <method name="getNodeUsage" return="java.lang.String"
- abstract="true" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[Get Max, Median, Min and Standard Deviation of DataNodes usage.
- @return the DataNode usage information, as a JSON string.]]>
- </doc>
- </method>
- <method name="getNameJournalStatus" return="java.lang.String"
- abstract="true" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[Get status information about the journals of the NN.
- @return the name journal status information, as a JSON string.]]>
- </doc>
- </method>
- <method name="getJournalTransactionInfo" return="java.lang.String"
- abstract="true" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[Get information about the transaction ID, including the last applied
- transaction ID and the most recent checkpoint's transaction ID]]>
- </doc>
- </method>
- <method name="getNNStarted" return="java.lang.String"
- abstract="true" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[Gets the NN start time
- @return the NN start time]]>
- </doc>
- </method>
- <method name="getCompileInfo" return="java.lang.String"
- abstract="true" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[Get the compilation information which contains date, user and branch
- @return the compilation information, as a JSON string.]]>
- </doc>
- </method>
- <method name="getCorruptFiles" return="java.lang.String"
- abstract="true" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[Get the list of corrupt files
- @return the list of corrupt files, as a JSON string.]]>
- </doc>
- </method>
- <method name="getDistinctVersionCount" return="int"
- abstract="true" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[Get the number of distinct versions of live datanodes
- @return the number of distinct versions of live datanodes]]>
- </doc>
- </method>
- <method name="getDistinctVersions" return="java.util.Map"
- abstract="true" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[Get the number of live datanodes for each distinct versions
- @return the number of live datanodes for each distinct versions]]>
- </doc>
- </method>
- <doc>
- <![CDATA[This is the JMX management interface for namenode information]]>
- </doc>
- </interface>
- <!-- end interface org.apache.hadoop.hdfs.server.namenode.NameNodeMXBean -->
- <!-- start interface org.apache.hadoop.hdfs.server.namenode.NameNodeStatusMXBean -->
- <interface name="NameNodeStatusMXBean" abstract="true"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <method name="getNNRole" return="java.lang.String"
- abstract="true" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[Gets the NameNode role.
- @return the NameNode role.]]>
- </doc>
- </method>
- <method name="getState" return="java.lang.String"
- abstract="true" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[Gets the NameNode state.
- @return the NameNode state.]]>
- </doc>
- </method>
- <method name="getHostAndPort" return="java.lang.String"
- abstract="true" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[Gets the host and port colon separated.
- @return host and port colon separated.]]>
- </doc>
- </method>
- <method name="isSecurityEnabled" return="boolean"
- abstract="true" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[Gets if security is enabled.
- @return true, if security is enabled.]]>
- </doc>
- </method>
- <method name="getLastHATransitionTime" return="long"
- abstract="true" native="false" synchronized="false"
- static="false" final="false" visibility="public"
- deprecated="not deprecated">
- <doc>
- <![CDATA[Gets the most recent HA transition time in milliseconds from the epoch.
- @return the most recent HA transition time in milliseconds from the epoch.]]>
- </doc>
- </method>
- <doc>
- <![CDATA[This is the JMX management interface for NameNode status information]]>
- </doc>
- </interface>
- <!-- end interface org.apache.hadoop.hdfs.server.namenode.NameNodeStatusMXBean -->
- </package>
- <package name="org.apache.hadoop.hdfs.server.namenode.ha">
- </package>
- <package name="org.apache.hadoop.hdfs.server.namenode.metrics">
- </package>
- <package name="org.apache.hadoop.hdfs.server.namenode.snapshot">
- </package>
- <package name="org.apache.hadoop.hdfs.server.namenode.top">
- </package>
- <package name="org.apache.hadoop.hdfs.server.namenode.top.metrics">
- </package>
- <package name="org.apache.hadoop.hdfs.server.namenode.top.window">
- </package>
- <package name="org.apache.hadoop.hdfs.server.namenode.web.resources">
- </package>
- <package name="org.apache.hadoop.hdfs.server.protocol">
- </package>
- <package name="org.apache.hadoop.hdfs.shortcircuit">
- </package>
- <package name="org.apache.hadoop.hdfs.tools">
- </package>
- <package name="org.apache.hadoop.hdfs.tools.offlineEditsViewer">
- </package>
- <package name="org.apache.hadoop.hdfs.tools.offlineImageViewer">
- </package>
- <package name="org.apache.hadoop.hdfs.tools.snapshot">
- </package>
- <package name="org.apache.hadoop.hdfs.util">
- </package>
- <package name="org.apache.hadoop.hdfs.web">
- </package>
- <package name="org.apache.hadoop.hdfs.web.resources">
- </package>
- </api>
|