File: //usr/local/lib/python3.10/test/__pycache__/test_tokenize.cpython-310.pyc
o
�i � @ s� d dl mZ d dlmZ d dlmZmZmZmZmZm Z m
Z
mZmZm
Z
mZmZmZmZmZ d dlmZmZ d dlZd dlmZ d dlmZmZ d dlmZmZ d dlmZ d d lm Z m!Z! d dl"Z"d dl#Z#d
d� Z$G dd
� d
e�Z%G dd� de%�Z&dd� Z'G dd� de�Z(G dd� de�Z)G dd� de�Z*G dd� de�Z+G dd� de�Z,G dd� de�Z-G dd� de�Z.G d d!� d!ej�Z/e0d"kr�e�1� dS dS )#� )�support)� os_helper)�tokenize� _tokenize�
untokenize�NUMBER�NAME�OP�STRING� ENDMARKER�ENCODING�tok_name�detect_encoding�open�Untokenizer�generate_tokens�NEWLINE)�BytesIO�StringION)�dedent)�TestCase�mock)�VALID_UNDERSCORE_LITERALS�INVALID_UNDERSCORE_LITERALS)�run_test_script�make_scriptc
C s� g }t |�� �}|d dv}| D ]2\}}}}} |tkr |S |r+|tkr+|d |kr+qt| }|�d|d�d|d�d|� d|� �� q|S )N���z
r � Z10� Z13)�len�
splitlinesr r r
�append)
Ztoken_generatorZ
source_string�resultZ num_linesZmissing_trailing_nl�type�token�start�end�line� r( �//usr/local/lib/python3.10/test/test_tokenize.py�stringify_tokens_from_source s �(r* c @ � e Zd Zdd� Zdd� Zdd� Zdd� Zd d
� Zdd� Zd
d� Z dd� Z
dd� Zdd� Zdd� Z
dd� Zdd� Zdd� Zdd� Zdd � Zd!d"� Zd#d$� Zd%d&� Zd'd(� Zd)S )*�TokenizeTestc C s<