neglect
1月前来过
全职 · 300/日  ·  6525/月
工作时间: 工作日07:00-22:00、周末07:00-23:00工作地点: 远程
服务企业: 1家累计提交: 1工时
联系方式:
********
********
********
聊一聊

使用APP扫码聊一聊

个人介绍

一名大三考研学生

学习计算机科学与技术专业

正在努力中

想要优秀就要拼命

已经学习了java python c语言 c++等多种语言

对于这方面甚是有兴趣,希望能在工作中学习到更多有用的知识

完善自己

让自己变得优秀

工作经历

  • 2020-07-23 -2020-08-23喜士客服务员

    兢兢业业,从不迟到,服务态度很好。能够及时处理好特殊情况。在这份工作中,锻炼了自己的心性格,能吃苦耐闹,服从安排,会从不同方向努力去完成上级分布的任务

教育经历

  • 2020-09-01 - 2022-01-11安徽三联学院计算机科学与技术本科

    在这个学校中,我十分遵守学校规章制度,成绩在班级中较为优越,曾经差点拿到奖学金,惜败惜败

技能

C++
Android
MySQL
0
1
2
3
4
5
0
1
2
3
4
5
作品
逻辑与

.text .global _start _start: mov r3,r3 ldr r0,=myarray mov r4,#0 loop: ldr r1,[r0],#4 add r4,r4,r1 cmp r1,#0 bne loop stop: b stop myarray: .word 0x20 .word 0x01 .word 0x01 .word 0x01 .word 0x76 .word 0x00 .end

0
2022-01-11 11:23
Python爬虫评论

import csv from lxml import etree import requests import json import time #输入头文件 headers = { 'User-Agent': 'Mozilla/5.0 (Windows NT 6.1; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/64.0.3282.140 Safari/537.36', } #将获得的评论写入一个csv文件 f1 = open('知乎评论.csv', 'w', encoding='utf-8') writer = csv.writer(f1) writer.writerow(['id', 'created_time', 'author', 'content']) def A(): i = 0 while True: #通过f12找到网址 url = 'https://www.zhihu.com/api/v4/questions/20270540/answers' \ '?include=data[*].is_normal,admin_closed_comment,reward_info,is_collapsed,annotation_action,' \ 'annotation_detail,collapse_reason,is_sticky,collapsed_by,suggest_edit,comment_count,can_comment,' \ 'content,editable_content,attachment,voteup_count,reshipment_settings,comment_permission,created_time,' \ 'updated_time,review_info,relevant_info,question,excerpt,is_labeled,paid_info,paid_info_content,relationship.is_authorized,is_author,' \ 'voting,is_thanked,is_nothelp,is_recognized;data[*].mark_infos[*].url;data[*].author.follower_count,vip_info,badge[*].topics;data[*].settings.table_of_content.enabled&limit=5&offset=5&platform=desktop' 'sort_by=default'.format(i)#format(i)用作换页 state=1 while state: try: res = requests.get(url, headers=headers) state=0 except: continue res.encoding = 'utf-8' #用json解析text js = json.loads(res.text) is_end = js['paging']['is_end']#筛选信息 for i in js['data']: l = list() id_1 = str(i['id']) l.append(id_1) l.append(time.strftime("%Y-%m-%d %H:%M:%S", time.localtime(i['created_time']))) l.append(i['author']['name']) l.append(''.join(etree.HTML(i['content']).xpath('//p//text()'))) writer.writerow(l) print(l) if i['admin_closed_comment'] == False and i['can_comment']['status'] and i['comment_count'] > 0: B(id_1) i += 5 print('打印到第{0}页'.format(int(i / 5))) if is_end: break time.sleep(2) def B(id_1): j = 0 while 1: url = 'https://www.zhihu.com/api/v4/answers/{0}/root_comments?order=normal&limit=20&offset={1}&status=open'.format(id_1, j) a=1 while a: try: res = requests.get(url, headers=headers) a=0 except: continue res.encoding = 'utf-8' jsC = json.loads(res.text) is_end = jsC['paging']['is_end'] for data in jsC['data']: l = list() ID_2 = str(id_1) + "_" + str(data['id']) l.append(ID_2) l.append(time.strftime("%Y-%m-%d %H:%M:%S", time.localtime(data['created_time']))) l.append(data['author']['member']['name']) l.append(''.join(etree.HTML(data['content']).xpath('//p//text()'))) writer.writerow(l) print(l) for J in data['child_comments']: l.clear() l.append(str(ID_2) + "_" + str(J['id'])) l.append(time.strftime("%Y-%m-%d %H:%M:%S", time.localtime(J['created_time']))) l.append(J['author']['member']['name']) l.append(''.join(etree.HTML(J['content']).xpath('//p//text()'))) writer.writerow(l) print(l) j += 20 if is_end: break time.sleep(2) A() f1.close()

0
2023-02-24 08:32
Python实现各种算法

for i in range(1, 10): for j in range(1, i+1): print('{}x{}={}\t'.format(j, i, i*j), end='') print()

0
2023-02-24 08:32
更新于: 2021-08-23 浏览: 238