Best Python code snippet using avocado_python
process.py
Source:process.py  
...9    # Writes to file for analysis10    def basics(self):11        print("PROCESSING BASIC STATISTICS - OUTPUT = results/basicstats.txt")12        total_tweets = "Total Tweets : " + str(self.stat_process.total_tweets())13        self.file_controller.append_one_line("results/basicstats.txt", total_tweets)14        tweets_min, group_min = self.stat_process.smallest_tweets_group()15        group_least_tweets = "Group with the least tweets : " + str(group_min) + ". Quantity : " + str(tweets_min)16        self.file_controller.append_one_line("results/basicstats.txt", group_least_tweets)17        tweets_max, group_max = self.stat_process.largest_tweets_group()18        group_most_tweets = "Group with the most tweets : " + str(group_max) + ". Quantity : " + str(tweets_max)19        self.file_controller.append_one_line("results/basicstats.txt", group_most_tweets)20        total_retweets = "Total Retweets : " + str(self.stat_process.total_retweets())21        self.file_controller.append_one_line("results/basicstats.txt", "\n" + total_retweets)22        tweets_min, group_min = self.stat_process.smallest_retweets_group()23        group_least_retweets = "Group with least retweets : " + str(group_min) + ". Quantity : " + str(tweets_min)24        self.file_controller.append_one_line("results/basicstats.txt", group_least_retweets)25        tweets_max, group_max = self.stat_process.largest_retweets_group()26        group_most_retweets = "Group with most retweets : " + str(group_max) + ". Quantity : " + str(tweets_max)27        self.file_controller.append_one_line("results/basicstats.txt", group_most_retweets)28        total_quotes = "Total Quotes : " + str(self.stat_process.total_quotes())29        self.file_controller.append_one_line("results/basicstats.txt", "\n" + total_quotes)30        tweets_min, group_min = self.stat_process.smallest_quote_group()31        group_least_quotes = "Group with least quotes : " + str(group_min) + ". Quantity : " + str(tweets_min)32        self.file_controller.append_one_line("results/basicstats.txt", group_least_quotes)33        tweets_max, group_max = self.stat_process.largest_quote_group()34        group_least_quotes = "Group with most quotes : " + str(group_max) + ". Quantity : " + str(tweets_max)35        self.file_controller.append_one_line("results/basicstats.txt", group_least_quotes)36        total_replies = "Total Replies : " + str(self.stat_process.total_replies())37        self.file_controller.append_one_line("results/basicstats.txt", "\n" + total_replies)38        tweets_min, group_min = self.stat_process.smallest_reply_group()39        group_least_replies = "Group with least replies : " + str(group_min) + ". Quantity : " + str(tweets_min)40        self.file_controller.append_one_line("results/basicstats.txt", group_least_replies)41        tweets_max, group_max = self.stat_process.largest_reply_group()42        group_least_replies = "Group with most replies : " + str(group_max) + ". Quantity : " + str(tweets_max)43        self.file_controller.append_one_line("results/basicstats.txt", group_least_replies)44    # Creates quotes network interaction map and processes centrality measures45    # Writes all data to a CSV file for further analysis46    # Writes user reable statistics to .txt file47    def quotes(self):48        print("PROCESSING QUOTE STATISTICS\nBasic Statistics output = results/quotesstats.txt\nAll statistics output = results/quotes.csv")49        network_process = interactiongraph.interactiongraph()50        self.file_controller.append_one_line("results/quotes.csv", "type, group, results")51        for i in range(20+1):52            print("Processing quotes for group " + str(i))53            if i == 0:54                retweets = network_process.quotes()55                group = "All"56                total_quotes = "Total Quotes : " + str(self.stat_process.total_quotes())57            else:58                retweets = network_process.quotes_group(i-1)59                group = i-160                total_quotes = "Total Quotes : " + str(self.stat_process.total_quote_group(group))61            self.file_controller.append_one_line("results/quotesstats.txt", "STATS FOR GROUP " + str(group))62            self.file_controller.append_one_line("results/quotesstats.txt", total_quotes)63            self.processing(retweets, group, "quotes")64    # Creates retweets network interaction map and processes centrality measures65    # Writes all data to a CSV file for further analysis66    # Writes user reable statistics to .txt file67    def retweets(self):68        print("PROCESSING RETWEETS STATISTICS\nBasic Statistics output = results/retweetsstats.txt\nAll statistics output = results/retweets.csv")69        network_process = interactiongraph.interactiongraph()70        self.file_controller.append_one_line("results/retweets.csv", "type, group, results")71        for i in range(20+1):72            print("Processing retweets for group " + str(i))73            if i == 0:74                retweets = network_process.retweets()75                group = "All"76                total_quotes = "Total Retweets : " + str(self.stat_process.total_retweets())77            else:78                retweets = network_process.retweets_group(i-1)79                group = i-180                total_quotes = "Total Retweets : " + str(self.stat_process.total_retweets_group(group))81            self.file_controller.append_one_line("results/retweetsstats.txt", "STATS FOR GROUP " + str(group))82            self.file_controller.append_one_line("results/quotesstats.txt", total_quotes)83            self.processing(retweets, group, "retweets")84    # Creates hashtags network interaction map and processes centrality measures85    # Writes all data to a CSV file for further analysis86    # Writes user reable statistics to .txt file87    def hashtags(self):88        print("PROCESSING HASHTAG STATISTICS\nBasic Statistics output = results/hashtagsstats.txt\nAll statistics output = results/hashtags.csv")89        network_process = interactiongraph.interactiongraph()90        self.file_controller.append_one_line("results/hashtags.csv", "type, group, results")91        for i in range(20+1):92            print("Processing hashtags for group " + str(i))93            if i == 0:94                retweets = network_process.hashtags_user('all')95                group = "All"96            else:97                retweets = network_process.retweets_group(i-1)98                group = i-199                self.processing(retweets, group, "hashtags")100    # Calculates all centrality measures for the given group (or all) and finds the groups which have the max and min of each101    def processing(self, retweets, group, name):102        centrality_process = centrality.centrality()103        in_degree_retweets, out_degree_retweets = centrality_process.in_out_centrality(retweets)104        in_degree_results = "IN DEGREE, " + str(group) + ", " + str(in_degree_retweets)105        self.file_controller.append_one_line("results/" + name + ".csv", in_degree_results)106        out_degree_results = "OUT DEGREE, " + str(group) + ", " + str(out_degree_retweets)107        self.file_controller.append_one_line("results/" + name + ".csv", out_degree_results)108        if not in_degree_retweets == None:109            in_degree_max = "In degree maximum for " + name + " from group " + str(group) + " is " + str(max(in_degree_retweets.values()))110            self.file_controller.append_one_line("results/" + name + "stats.txt", in_degree_max)111            out_degree_max = "Out degree maximum for " + name + " from group " + str(group) + " is " + str(max(out_degree_retweets.values()))112            self.file_controller.append_one_line("results/" + name + "stats.txt", out_degree_max)113            in_degree_min = "In degree minimum for " + name + " from group " + str(group) + " is " + str(min(in_degree_retweets.values()))114            self.file_controller.append_one_line("results/" + name + "stats.txt", in_degree_min)115            out_degree_min = "Out degree minimum for " + name + " from group " + str(group) + " is " + str(min(out_degree_retweets.values()))116            self.file_controller.append_one_line("results/" + name + "stats.txt", out_degree_min)117        betweenness = centrality_process.betweenness(retweets)118        betweenness_results = "BETWEENNESS," + str(group) + ", " + str(betweenness)119        self.file_controller.append_one_line("results/" + name + ".csv", betweenness_results)120        if not len(betweenness) == 0:121            betweenness_max = "Betweenness degree maximum for " + name + " from group " + str(group) + " is " + str(max(betweenness.values()))122            self.file_controller.append_one_line("results/" + name + "stats.txt", betweenness_max)123            betweenness_min = "Betweenness degree minimum for " + name + " from group " + str(group) + " is " + str(min(betweenness.values()))124            self.file_controller.append_one_line("results/" + name + "stats.txt", betweenness_min)125        closeness = centrality_process.closeness(retweets)126        closeness_results = "CLOSENESS, " + str(group) + ", " + str(closeness)127        self.file_controller.append_one_line("results/" + name + ".csv", closeness_results)128        if not len(closeness) == 0:129            closeness_max = "Closeness degree maximum for " + name + " from group " + str(group) + " is " + str(max(closeness.values()))130            self.file_controller.append_one_line("results/" + name + "stats.txt", closeness_max)131            closeness_min = "Closeness degree minimum for " + name + " from group " + str(group) + " is " + str(min(closeness.values()))132            self.file_controller.append_one_line("results/" + name + "stats.txt", closeness_min)133        eigenvector = centrality_process.eigenvector(retweets)134        eigenvector_results = "EIGENVECTOR, " + str(group) + ", " + str(eigenvector)135        self.file_controller.append_one_line("results/" + name + ".csv", eigenvector_results)136        if eigenvector:137            eigen_max = "Eigenvector degree maximum for " + name + " from group " + str(group) + " is " + str(max(eigenvector.values()))138            self.file_controller.append_one_line("results/" + name + "stats.txt", eigen_max)139            eigen_min = "Eigenvector degree minimum for " + name + " from group " + str(group) + " is " + str(min(eigenvector.values()))140            self.file_controller.append_one_line("results/" + name + "stats.txt", eigen_min)...base_accessor.py
Source:base_accessor.py  
...88        :return:89        """90        occurrence = self.get_setting_occurrence(name)91        return True if occurrence else False92    def append_one_line(self, line):93        self.lines.append(line + "\n")94    def add_setting(self, name, value, add_comment):95        if add_comment:96            self.append_one_line(self.file_config.generate_setting_comment(name))97        self.append_one_line(self.file_config.generate_setting_line(name, value))98    def update_setting(self, name, value, occurrence=None):99        if not occurrence:100            occurrence = self.get_setting_occurrence(name)101        if not occurrence:102            return SettingNotFoundException("can not find setting of %s while updating" % name, name)103        line = self.file_config.generate_setting_line(name, value)...classes.py
Source:classes.py  
...41  42  def __len__(self) -> int:43    return len(self.lines)44  45  def append_one_line(self, l:Line) -> None:46    if len(self.lines) == 0:47      # self.lines.append = [l]48      self.lines = np.append(self.lines, l)49      self.maxH = l.top50      self.maxW = l.index51      self.minH = l.down52      self.minW = l.index53      return54    else:55      self.lines = np.append(self.lines, l)56      # self.lines.append(l)57    # self.lines = sorted(self.lines)58    59    self.maxH = max(l.top, self.maxH)60    self.maxW = max(l.index, self.maxW)61    self.minH = min(l.down, self.minH)62    self.minW = min(l.index, self.minW)63  def sort(self) -> None:64    self.lines = np.sort(self.lines)65  def smooth(self):66    # newlines = [self.lines[0]]67    newlines = np.empty(0, dtype=line_np_type)68    newlines = np.append(newlines, self.lines[0])69    for l in self.lines[1:]:70      if newlines[-1].index != l.index:71        newlines = np.append(newlines, l)72        continue73      newlines[-1].top  = max(newlines[-1].top,  l.top)74      newlines[-1].down = min(newlines[-1].down, l.down)75    self.lines = newlines76  def __add__(self, other):77    for i in other:78      self.append_one_line(i)79      80    self.lines = np.sort(self.lines)81    return self82  def __repr__(self):...Learn to execute automation testing from scratch with LambdaTest Learning Hub. Right from setting up the prerequisites to run your first automation test, to following best practices and diving deeper into advanced test scenarios. LambdaTest Learning Hubs compile a list of step-by-step guides to help you be proficient with different test automation frameworks i.e. Selenium, Cypress, TestNG etc.
You could also refer to video tutorials over LambdaTest YouTube channel to get step by step demonstration from industry experts.
Get 100 minutes of automation test minutes FREE!!
