2022-01-12 15:07:34 +01:00
|
|
|
import sys
|
|
|
|
import os
|
2020-09-28 12:16:57 +02:00
|
|
|
import zlib
|
2021-12-29 10:41:44 +01:00
|
|
|
import functools
|
|
|
|
from typing import Any, Callable, IO, Iterable, Optional, Tuple, Union, TYPE_CHECKING
|
2020-09-28 12:16:57 +02:00
|
|
|
from urllib.parse import urlencode
|
2020-08-15 17:50:00 +02:00
|
|
|
|
2020-09-28 12:16:57 +02:00
|
|
|
import requests
|
2020-09-25 13:44:28 +02:00
|
|
|
from requests.utils import super_len
|
2021-12-17 09:00:22 +01:00
|
|
|
|
|
|
|
if TYPE_CHECKING:
|
|
|
|
from requests_toolbelt import MultipartEncoder
|
2020-08-15 17:50:00 +02:00
|
|
|
|
2022-01-12 15:07:34 +01:00
|
|
|
from .context import Environment
|
2021-05-05 14:13:39 +02:00
|
|
|
from .cli.dicts import MultipartRequestDataDict, RequestDataDict
|
2022-01-12 15:07:34 +01:00
|
|
|
from .compat import is_windows
|
2020-08-19 10:22:42 +02:00
|
|
|
|
2020-08-15 17:50:00 +02:00
|
|
|
|
2021-12-29 10:41:44 +01:00
|
|
|
class ChunkedStream:
|
|
|
|
def __iter__(self) -> Iterable[Union[str, bytes]]:
|
|
|
|
raise NotImplementedError
|
|
|
|
|
|
|
|
|
|
|
|
class ChunkedUploadStream(ChunkedStream):
|
2020-09-28 12:16:57 +02:00
|
|
|
def __init__(self, stream: Iterable, callback: Callable):
|
|
|
|
self.callback = callback
|
|
|
|
self.stream = stream
|
|
|
|
|
|
|
|
def __iter__(self) -> Iterable[Union[str, bytes]]:
|
|
|
|
for chunk in self.stream:
|
|
|
|
self.callback(chunk)
|
|
|
|
yield chunk
|
|
|
|
|
|
|
|
|
2021-12-29 10:41:44 +01:00
|
|
|
class ChunkedMultipartUploadStream(ChunkedStream):
|
2020-09-28 12:16:57 +02:00
|
|
|
chunk_size = 100 * 1024
|
|
|
|
|
2021-12-17 09:00:22 +01:00
|
|
|
def __init__(self, encoder: 'MultipartEncoder'):
|
2020-09-28 12:16:57 +02:00
|
|
|
self.encoder = encoder
|
|
|
|
|
|
|
|
def __iter__(self) -> Iterable[Union[str, bytes]]:
|
|
|
|
while True:
|
|
|
|
chunk = self.encoder.read(self.chunk_size)
|
|
|
|
if not chunk:
|
|
|
|
break
|
|
|
|
yield chunk
|
|
|
|
|
|
|
|
|
2021-12-29 10:41:44 +01:00
|
|
|
def as_bytes(data: Union[str, bytes]) -> bytes:
|
|
|
|
if isinstance(data, str):
|
|
|
|
return data.encode()
|
|
|
|
else:
|
|
|
|
return data
|
|
|
|
|
|
|
|
|
|
|
|
CallbackT = Callable[[bytes], bytes]
|
|
|
|
|
|
|
|
|
|
|
|
def _wrap_function_with_callback(
|
|
|
|
func: Callable[..., Any],
|
|
|
|
callback: CallbackT
|
|
|
|
) -> Callable[..., Any]:
|
|
|
|
@functools.wraps(func)
|
|
|
|
def wrapped(*args, **kwargs):
|
|
|
|
chunk = func(*args, **kwargs)
|
|
|
|
callback(chunk)
|
|
|
|
return chunk
|
|
|
|
return wrapped
|
|
|
|
|
|
|
|
|
2022-01-12 15:07:34 +01:00
|
|
|
def is_stdin(file: IO) -> bool:
|
|
|
|
try:
|
|
|
|
file_no = file.fileno()
|
|
|
|
except Exception:
|
|
|
|
return False
|
|
|
|
else:
|
|
|
|
return file_no == sys.stdin.fileno()
|
|
|
|
|
|
|
|
|
|
|
|
READ_THRESHOLD = float(os.getenv("HTTPIE_STDIN_READ_WARN_THRESHOLD", 10.0))
|
|
|
|
|
|
|
|
|
|
|
|
def observe_stdin_for_data_thread(env: Environment, file: IO) -> None:
|
|
|
|
# Windows unfortunately does not support select() operation
|
|
|
|
# on regular files, like stdin in our use case.
|
|
|
|
# https://docs.python.org/3/library/select.html#select.select
|
|
|
|
if is_windows:
|
|
|
|
return None
|
|
|
|
|
|
|
|
# If the user configures READ_THRESHOLD to be 0, then
|
|
|
|
# disable this warning.
|
|
|
|
if READ_THRESHOLD == 0:
|
|
|
|
return None
|
|
|
|
|
|
|
|
import select
|
|
|
|
import threading
|
|
|
|
|
|
|
|
def worker():
|
|
|
|
can_read, _, _ = select.select([file], [], [], READ_THRESHOLD)
|
|
|
|
if not can_read:
|
|
|
|
env.stderr.write(
|
|
|
|
f'> warning: no stdin data read in {READ_THRESHOLD}s '
|
|
|
|
f'(perhaps you want to --ignore-stdin)\n'
|
|
|
|
f'> See: https://httpie.io/docs/cli/best-practices\n'
|
|
|
|
)
|
|
|
|
|
|
|
|
thread = threading.Thread(
|
|
|
|
target=worker
|
|
|
|
)
|
|
|
|
thread.start()
|
|
|
|
|
|
|
|
|
2021-12-29 10:41:44 +01:00
|
|
|
def _prepare_file_for_upload(
|
2022-01-12 15:07:34 +01:00
|
|
|
env: Environment,
|
2021-12-29 10:41:44 +01:00
|
|
|
file: Union[IO, 'MultipartEncoder'],
|
|
|
|
callback: CallbackT,
|
|
|
|
chunked: bool = False,
|
|
|
|
content_length_header_value: Optional[int] = None,
|
|
|
|
) -> Union[bytes, IO, ChunkedStream]:
|
|
|
|
if not super_len(file):
|
2022-01-12 15:07:34 +01:00
|
|
|
if is_stdin(file):
|
|
|
|
observe_stdin_for_data_thread(env, file)
|
2021-12-29 10:41:44 +01:00
|
|
|
# Zero-length -> assume stdin.
|
|
|
|
if content_length_header_value is None and not chunked:
|
|
|
|
# Read the whole stdin to determine `Content-Length`.
|
|
|
|
#
|
|
|
|
# TODO: Instead of opt-in --chunked, consider making
|
|
|
|
# `Transfer-Encoding: chunked` for STDIN opt-out via
|
|
|
|
# something like --no-chunked.
|
|
|
|
# This would be backwards-incompatible so wait until v3.0.0.
|
|
|
|
#
|
|
|
|
file = as_bytes(file.read())
|
|
|
|
else:
|
|
|
|
file.read = _wrap_function_with_callback(
|
|
|
|
file.read,
|
|
|
|
callback
|
|
|
|
)
|
|
|
|
|
|
|
|
if chunked:
|
|
|
|
from requests_toolbelt import MultipartEncoder
|
|
|
|
if isinstance(file, MultipartEncoder):
|
|
|
|
return ChunkedMultipartUploadStream(
|
|
|
|
encoder=file,
|
|
|
|
)
|
|
|
|
else:
|
|
|
|
return ChunkedUploadStream(
|
|
|
|
stream=file,
|
|
|
|
callback=callback,
|
|
|
|
)
|
|
|
|
else:
|
|
|
|
return file
|
2020-09-28 16:22:34 +02:00
|
|
|
|
|
|
|
|
2021-12-29 10:41:44 +01:00
|
|
|
def prepare_request_body(
|
2022-01-12 15:07:34 +01:00
|
|
|
env: Environment,
|
2021-12-29 10:41:44 +01:00
|
|
|
raw_body: Union[str, bytes, IO, 'MultipartEncoder', RequestDataDict],
|
|
|
|
body_read_callback: CallbackT,
|
|
|
|
offline: bool = False,
|
|
|
|
chunked: bool = False,
|
|
|
|
content_length_header_value: Optional[int] = None,
|
|
|
|
) -> Union[bytes, IO, 'MultipartEncoder', ChunkedStream]:
|
|
|
|
is_file_like = hasattr(raw_body, 'read')
|
|
|
|
if isinstance(raw_body, (bytes, str)):
|
|
|
|
body = as_bytes(raw_body)
|
|
|
|
elif isinstance(raw_body, RequestDataDict):
|
|
|
|
body = as_bytes(urlencode(raw_body, doseq=True))
|
|
|
|
else:
|
|
|
|
body = raw_body
|
2020-09-28 12:16:57 +02:00
|
|
|
|
2020-09-28 16:22:34 +02:00
|
|
|
if offline:
|
|
|
|
if is_file_like:
|
2021-12-29 10:41:44 +01:00
|
|
|
return as_bytes(raw_body.read())
|
2020-09-28 12:16:57 +02:00
|
|
|
else:
|
2021-12-29 10:41:44 +01:00
|
|
|
return body
|
|
|
|
|
|
|
|
if is_file_like:
|
|
|
|
return _prepare_file_for_upload(
|
2022-01-12 15:07:34 +01:00
|
|
|
env,
|
2021-12-29 10:41:44 +01:00
|
|
|
body,
|
|
|
|
chunked=chunked,
|
|
|
|
callback=body_read_callback,
|
|
|
|
content_length_header_value=content_length_header_value
|
|
|
|
)
|
|
|
|
elif chunked:
|
|
|
|
return ChunkedUploadStream(
|
|
|
|
stream=iter([body]),
|
|
|
|
callback=body_read_callback
|
|
|
|
)
|
|
|
|
else:
|
|
|
|
return body
|
2020-08-15 17:50:00 +02:00
|
|
|
|
|
|
|
|
2020-08-19 10:22:42 +02:00
|
|
|
def get_multipart_data_and_content_type(
|
2020-09-28 12:16:57 +02:00
|
|
|
data: MultipartRequestDataDict,
|
2020-08-19 10:22:42 +02:00
|
|
|
boundary: str = None,
|
|
|
|
content_type: str = None,
|
2021-12-17 09:00:22 +01:00
|
|
|
) -> Tuple['MultipartEncoder', str]:
|
|
|
|
from requests_toolbelt import MultipartEncoder
|
|
|
|
|
2020-08-19 10:22:42 +02:00
|
|
|
encoder = MultipartEncoder(
|
2020-09-28 12:16:57 +02:00
|
|
|
fields=data.items(),
|
2020-08-19 10:22:42 +02:00
|
|
|
boundary=boundary,
|
|
|
|
)
|
|
|
|
if content_type:
|
2020-09-28 12:16:57 +02:00
|
|
|
content_type = content_type.strip()
|
2020-08-19 10:22:42 +02:00
|
|
|
if 'boundary=' not in content_type:
|
|
|
|
content_type = f'{content_type}; boundary={encoder.boundary_value}'
|
|
|
|
else:
|
|
|
|
content_type = encoder.content_type
|
|
|
|
|
2020-09-28 12:16:57 +02:00
|
|
|
data = encoder
|
2020-08-15 17:50:00 +02:00
|
|
|
return data, content_type
|
2020-09-25 13:44:28 +02:00
|
|
|
|
|
|
|
|
2020-09-28 12:16:57 +02:00
|
|
|
def compress_request(
|
|
|
|
request: requests.PreparedRequest,
|
|
|
|
always: bool,
|
|
|
|
):
|
|
|
|
deflater = zlib.compressobj()
|
|
|
|
if isinstance(request.body, str):
|
|
|
|
body_bytes = request.body.encode()
|
|
|
|
elif hasattr(request.body, 'read'):
|
|
|
|
body_bytes = request.body.read()
|
|
|
|
else:
|
|
|
|
body_bytes = request.body
|
|
|
|
deflated_data = deflater.compress(body_bytes)
|
|
|
|
deflated_data += deflater.flush()
|
|
|
|
is_economical = len(deflated_data) < len(body_bytes)
|
|
|
|
if is_economical or always:
|
|
|
|
request.body = deflated_data
|
|
|
|
request.headers['Content-Encoding'] = 'deflate'
|
|
|
|
request.headers['Content-Length'] = str(len(deflated_data))
|