🔨 cleanup
This commit is contained in:
		
							parent
							
								
									850671b9f1
								
							
						
					
					
						commit
						ba6edd5634
					
				
					 2 changed files with 14 additions and 36 deletions
				
			
		|  | @ -182,11 +182,8 @@ class parser | |||
|                             return false; | ||||
|                         } | ||||
| 
 | ||||
|                         // read next token
 | ||||
|                         get_token(); | ||||
| 
 | ||||
|                         // closing } -> we are done
 | ||||
|                         if (last_token == token_type::end_object) | ||||
|                         if (get_token() == token_type::end_object) | ||||
|                         { | ||||
|                             if (JSON_UNLIKELY(not sax->end_object())) | ||||
|                             { | ||||
|  | @ -211,8 +208,7 @@ class parser | |||
|                         } | ||||
| 
 | ||||
|                         // parse separator (:)
 | ||||
|                         get_token(); | ||||
|                         if (JSON_UNLIKELY(last_token != token_type::name_separator)) | ||||
|                         if (JSON_UNLIKELY(get_token() != token_type::name_separator)) | ||||
|                         { | ||||
|                             return sax->parse_error(m_lexer.get_position(), | ||||
|                                                     m_lexer.get_token_string(), | ||||
|  | @ -234,11 +230,8 @@ class parser | |||
|                             return false; | ||||
|                         } | ||||
| 
 | ||||
|                         // read next token
 | ||||
|                         get_token(); | ||||
| 
 | ||||
|                         // closing ] -> we are done
 | ||||
|                         if (last_token == token_type::end_array) | ||||
|                         if (get_token() == token_type::end_array) | ||||
|                         { | ||||
|                             if (JSON_UNLIKELY(not sax->end_array())) | ||||
|                             { | ||||
|  | @ -357,11 +350,10 @@ class parser | |||
|             } | ||||
|             else | ||||
|             { | ||||
|                 get_token(); | ||||
|                 if (states.back())  // array
 | ||||
|                 { | ||||
|                     // comma -> next value
 | ||||
|                     if (last_token == token_type::value_separator) | ||||
|                     if (get_token() == token_type::value_separator) | ||||
|                     { | ||||
|                         // parse a new value
 | ||||
|                         get_token(); | ||||
|  | @ -395,12 +387,10 @@ class parser | |||
|                 else  // object
 | ||||
|                 { | ||||
|                     // comma -> next value
 | ||||
|                     if (last_token == token_type::value_separator) | ||||
|                     if (get_token() == token_type::value_separator) | ||||
|                     { | ||||
|                         get_token(); | ||||
| 
 | ||||
|                         // parse key
 | ||||
|                         if (JSON_UNLIKELY(last_token != token_type::value_string)) | ||||
|                         if (JSON_UNLIKELY(get_token() != token_type::value_string)) | ||||
|                         { | ||||
|                             return sax->parse_error(m_lexer.get_position(), | ||||
|                                                     m_lexer.get_token_string(), | ||||
|  | @ -415,8 +405,7 @@ class parser | |||
|                         } | ||||
| 
 | ||||
|                         // parse separator (:)
 | ||||
|                         get_token(); | ||||
|                         if (JSON_UNLIKELY(last_token != token_type::name_separator)) | ||||
|                         if (JSON_UNLIKELY(get_token() != token_type::name_separator)) | ||||
|                         { | ||||
|                             return sax->parse_error(m_lexer.get_position(), | ||||
|                                                     m_lexer.get_token_string(), | ||||
|  |  | |||
		Loading…
	
	Add table
		Add a link
		
	
		Reference in a new issue