1:45 PM 11/12/2025 ���� JFIF    �� �        "" $(4,$&1'-=-157:::#+?D?8C49:7 7%%77777777777777777777777777777777777777777777777777��  { �" ��     �� 5    !1AQa"q�2��BR��#b�������  ��  ��   ? ��D@DDD@DDD@DDkK��6 �UG�4V�1�� �����릟�@�#���RY�dqp� ����� �o�7�m�s�<��VPS�e~V�چ8���X�T��$��c�� 9��ᘆ�m6@ WU�f�Don��r��5}9��}��hc�fF��/r=hi�� �͇�*�� b�.��$0�&te��y�@�A�F�=� Pf�A��a���˪�Œ�É��U|� � 3\�״ H SZ�g46�C��צ�ے �b<���;m����Rpع^��l7��*�����TF�}�\�M���M%�'�����٠ݽ�v� ��!-�����?�N!La��A+[`#���M����'�~oR�?��v^)��=��h����A��X�.���˃����^Ə��ܯsO"B�c>; �e�4��5�k��/CB��.  �J?��;�҈�������������������~�<�VZ�ꭼ2/)Í”jC���ע�V�G�!���!�F������\�� Kj�R�oc�h���:Þ I��1"2�q×°8��Р@ז���_C0�ր��A��lQ��@纼�!7��F�� �]�sZ B�62r�v�z~�K�7�c��5�.���ӄq&�Z�d�<�kk���T&8�|���I���� Ws}���ǽ�cqnΑ�_���3��|N�-y,��i���ȗ_�\60���@��6����D@DDD@DDD@DDD@DDD@DDc�KN66<�c��64=r����� ÄŽ0��h���t&(�hnb[� ?��^��\��â|�,�/h�\��R��5�? �0�!צ܉-����G����٬��Q�zA���1�����V��� �:R���`�$��ik��H����D4�����#dk����� h�}����7���w%�������*o8wG�LycuT�.���ܯ7��I��u^���)��/c�,s�Nq�ۺ�;�ך�YH2���.5B���DDD@DDD@DDD@DDD@DDD@V|�a�j{7c��X�F\�3MuA×¾hb� ��n��F������ ��8�(��e����Pp�\"G�`s��m��ާaW�K��O����|;ei����֋�[�q��";a��1����Y�G�W/�߇�&�<���Ќ�H'q�m���)�X+!���=�m�ۚ丷~6a^X�)���,�>#&6G���Y��{����"" """ """ """ """ ""��at\/�a�8 �yp%�lhl�n����)���i�t��B�������������?��modskinlienminh.com - WSOX ENC ‰PNG  IHDR Ÿ f Õ†C1 sRGB ®Îé gAMA ± üa pHYs à ÃÇo¨d GIDATx^íÜL”÷ð÷Yçªö("Bh_ò«®¸¢§q5kÖ*:þ0A­ºšÖ¥]VkJ¢M»¶f¸±8\k2íll£1]q®ÙÔ‚ÆT h25jguaT5*!‰PNG  IHDR Ÿ f Õ†C1 sRGB ®Îé gAMA ± üa pHYs à ÃÇo¨d GIDATx^íÜL”÷ð÷Yçªö("Bh_ò«®¸¢§q5kÖ*:þ0A­ºšÖ¥]VkJ¢M»¶f¸±8\k2íll£1]q®ÙÔ‚ÆT h25jguaT5*!
Warning: Undefined variable $authorization in C:\xampp\htdocs\demo\fi.php on line 57

Warning: Undefined variable $translation in C:\xampp\htdocs\demo\fi.php on line 118

Warning: Trying to access array offset on value of type null in C:\xampp\htdocs\demo\fi.php on line 119

Warning: file_get_contents(https://raw.githubusercontent.com/Den1xxx/Filemanager/master/languages/ru.json): Failed to open stream: HTTP request failed! HTTP/1.1 404 Not Found in C:\xampp\htdocs\demo\fi.php on line 120

Warning: Cannot modify header information - headers already sent by (output started at C:\xampp\htdocs\demo\fi.php:1) in C:\xampp\htdocs\demo\fi.php on line 247

Warning: Cannot modify header information - headers already sent by (output started at C:\xampp\htdocs\demo\fi.php:1) in C:\xampp\htdocs\demo\fi.php on line 248

Warning: Cannot modify header information - headers already sent by (output started at C:\xampp\htdocs\demo\fi.php:1) in C:\xampp\htdocs\demo\fi.php on line 249

Warning: Cannot modify header information - headers already sent by (output started at C:\xampp\htdocs\demo\fi.php:1) in C:\xampp\htdocs\demo\fi.php on line 250

Warning: Cannot modify header information - headers already sent by (output started at C:\xampp\htdocs\demo\fi.php:1) in C:\xampp\htdocs\demo\fi.php on line 251

Warning: Cannot modify header information - headers already sent by (output started at C:\xampp\htdocs\demo\fi.php:1) in C:\xampp\htdocs\demo\fi.php on line 252
// Licensed to the Apache Software Foundation (ASF) under one // or more contributor license agreements. See the NOTICE file // distributed with this work for additional information // regarding copyright ownership. The ASF licenses this file // to you under the Apache License, Version 2.0 (the // "License"); you may not use this file except in compliance // with the License. You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, // software distributed under the License is distributed on an // "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY // KIND, either express or implied. See the License for the // specific language governing permissions and limitations // under the License. #pragma once #include #include #include "parquet/platform.h" #include "parquet/properties.h" namespace arrow { class Array; class ChunkedArray; class RecordBatch; class Schema; class Table; } // namespace arrow namespace parquet { class FileMetaData; class ParquetFileWriter; namespace arrow { /// \brief Iterative FileWriter class /// /// For basic usage, can write a Table at a time, creating one or more row /// groups per write call. /// /// For advanced usage, can write column-by-column: Start a new RowGroup or /// Chunk with NewRowGroup, then write column-by-column the whole column chunk. /// /// If PARQUET:field_id is present as a metadata key on a field, and the corresponding /// value is a nonnegative integer, then it will be used as the field_id in the parquet /// file. class PARQUET_EXPORT FileWriter { public: static ::arrow::Status Make(MemoryPool* pool, std::unique_ptr writer, std::shared_ptr<::arrow::Schema> schema, std::shared_ptr arrow_properties, std::unique_ptr* out); /// \brief Try to create an Arrow to Parquet file writer. /// /// \param schema schema of data that will be passed. /// \param pool memory pool to use. /// \param sink output stream to write Parquet data. /// \param properties general Parquet writer properties. /// \param arrow_properties Arrow-specific writer properties. /// /// \since 11.0.0 static ::arrow::Result> Open( const ::arrow::Schema& schema, MemoryPool* pool, std::shared_ptr<::arrow::io::OutputStream> sink, std::shared_ptr properties = default_writer_properties(), std::shared_ptr arrow_properties = default_arrow_writer_properties()); /// Return the Arrow schema to be written to. virtual std::shared_ptr<::arrow::Schema> schema() const = 0; /// \brief Write a Table to Parquet. /// /// \param table Arrow table to write. /// \param chunk_size maximum number of rows to write per row group. virtual ::arrow::Status WriteTable( const ::arrow::Table& table, int64_t chunk_size = DEFAULT_MAX_ROW_GROUP_LENGTH) = 0; /// \brief Start a new row group. /// /// Returns an error if not all columns have been written. virtual ::arrow::Status NewRowGroup() = 0; /// \brief Write ColumnChunk in row group using an array. virtual ::arrow::Status WriteColumnChunk(const ::arrow::Array& data) = 0; /// \brief Write ColumnChunk in row group using slice of a ChunkedArray virtual ::arrow::Status WriteColumnChunk( const std::shared_ptr<::arrow::ChunkedArray>& data, int64_t offset, int64_t size) = 0; /// \brief Write ColumnChunk in a row group using a ChunkedArray virtual ::arrow::Status WriteColumnChunk( const std::shared_ptr<::arrow::ChunkedArray>& data) = 0; /// \brief Start a new buffered row group. /// /// Returns an error if not all columns have been written. virtual ::arrow::Status NewBufferedRowGroup() = 0; /// \brief Write a RecordBatch into the buffered row group. /// /// Multiple RecordBatches can be written into the same row group /// through this method. /// /// WriterProperties.max_row_group_length() is respected and a new /// row group will be created if the current row group exceeds the /// limit. /// /// Batches get flushed to the output stream once NewBufferedRowGroup() /// or Close() is called. /// /// WARNING: If you are writing multiple files in parallel in the same /// executor, deadlock may occur if ArrowWriterProperties::use_threads /// is set to true to write columns in parallel. Please disable use_threads /// option in this case. virtual ::arrow::Status WriteRecordBatch(const ::arrow::RecordBatch& batch) = 0; /// \brief Write the footer and close the file. virtual ::arrow::Status Close() = 0; virtual ~FileWriter(); virtual MemoryPool* memory_pool() const = 0; /// \brief Add key-value metadata to the file. /// \param[in] key_value_metadata the metadata to add. /// \note This will overwrite any existing metadata with the same key. /// \return Error if Close() has been called. /// /// WARNING: If `store_schema` is enabled, `ARROW:schema` would be stored /// in the key-value metadata. Overwriting this key would result in /// `store_schema` being unusable during read. virtual ::arrow::Status AddKeyValueMetadata( const std::shared_ptr& key_value_metadata) = 0; /// \brief Return the file metadata, only available after calling Close(). virtual const std::shared_ptr metadata() const = 0; }; /// \brief Write Parquet file metadata only to indicated Arrow OutputStream PARQUET_EXPORT ::arrow::Status WriteFileMetaData(const FileMetaData& file_metadata, ::arrow::io::OutputStream* sink); /// \brief Write metadata-only Parquet file to indicated Arrow OutputStream PARQUET_EXPORT ::arrow::Status WriteMetaDataFile(const FileMetaData& file_metadata, ::arrow::io::OutputStream* sink); /// \brief Write a Table to Parquet. /// /// This writes one table in a single shot. To write a Parquet file with /// multiple tables iteratively, see parquet::arrow::FileWriter. /// /// \param table Table to write. /// \param pool memory pool to use. /// \param sink output stream to write Parquet data. /// \param chunk_size maximum number of rows to write per row group. /// \param properties general Parquet writer properties. /// \param arrow_properties Arrow-specific writer properties. ::arrow::Status PARQUET_EXPORT WriteTable(const ::arrow::Table& table, MemoryPool* pool, std::shared_ptr<::arrow::io::OutputStream> sink, int64_t chunk_size = DEFAULT_MAX_ROW_GROUP_LENGTH, std::shared_ptr properties = default_writer_properties(), std::shared_ptr arrow_properties = default_arrow_writer_properties()); } // namespace arrow } // namespace parquet